[ 451.507925] env[62569]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62569) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 451.508296] env[62569]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62569) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 451.508376] env[62569]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62569) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 451.508669] env[62569]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 451.602302] env[62569]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62569) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 451.612377] env[62569]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=62569) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 452.213361] env[62569]: INFO nova.virt.driver [None req-f01e0543-7b4c-43bb-a373-35541c37d100 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 452.285073] env[62569]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 452.285274] env[62569]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 452.285327] env[62569]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62569) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 455.396082] env[62569]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-ead52ab2-91e4-474a-b1d1-2ce1a579f83a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 455.413102] env[62569]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62569) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 455.413280] env[62569]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-60ef6506-9923-44dd-908d-84b75f0b0b05 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 455.448518] env[62569]: INFO oslo_vmware.api [-] Successfully established new session; session ID is cb87c. [ 455.448716] env[62569]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.163s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 455.449283] env[62569]: INFO nova.virt.vmwareapi.driver [None req-f01e0543-7b4c-43bb-a373-35541c37d100 None None] VMware vCenter version: 7.0.3 [ 455.452698] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eb420fc-b6d2-479b-9227-deaf9b78cf42 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 455.469717] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bc4675e-acb9-4242-b988-8573f4eefcd8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 455.475476] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9e7e1b7-1e44-42f3-934e-6e9991b908ca {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 455.482048] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac1f6dd-fb90-4659-8f94-a9564ce04ff4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 455.494983] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-063cc02c-cdc7-4455-b9eb-fd5ee48729ee {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 455.500766] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53fae00f-e4d3-42f5-8c89-c11e2d3f5de9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 455.531221] env[62569]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-68a9ae6f-3b8d-4541-88c3-30bf0bb59956 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 455.535944] env[62569]: DEBUG nova.virt.vmwareapi.driver [None req-f01e0543-7b4c-43bb-a373-35541c37d100 None None] Extension org.openstack.compute already exists. {{(pid=62569) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:227}} [ 455.538556] env[62569]: INFO nova.compute.provider_config [None req-f01e0543-7b4c-43bb-a373-35541c37d100 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 456.042219] env[62569]: DEBUG nova.context [None req-f01e0543-7b4c-43bb-a373-35541c37d100 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),0649a166-844b-46c9-9b27-cb9b95b9bc15(cell1) {{(pid=62569) load_cells /opt/stack/nova/nova/context.py:464}} [ 456.044266] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 456.044488] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 456.045180] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 456.045615] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] Acquiring lock "0649a166-844b-46c9-9b27-cb9b95b9bc15" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 456.045815] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] Lock "0649a166-844b-46c9-9b27-cb9b95b9bc15" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 456.046941] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] Lock "0649a166-844b-46c9-9b27-cb9b95b9bc15" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 456.067647] env[62569]: INFO dbcounter [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] Registered counter for database nova_cell0 [ 456.076068] env[62569]: INFO dbcounter [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] Registered counter for database nova_cell1 [ 456.079280] env[62569]: DEBUG oslo_db.sqlalchemy.engines [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62569) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 456.079696] env[62569]: DEBUG oslo_db.sqlalchemy.engines [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62569) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 456.084379] env[62569]: ERROR nova.db.main.api [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 456.084379] env[62569]: result = function(*args, **kwargs) [ 456.084379] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 456.084379] env[62569]: return func(*args, **kwargs) [ 456.084379] env[62569]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 456.084379] env[62569]: result = fn(*args, **kwargs) [ 456.084379] env[62569]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 456.084379] env[62569]: return f(*args, **kwargs) [ 456.084379] env[62569]: File "/opt/stack/nova/nova/objects/service.py", line 556, in _db_service_get_minimum_version [ 456.084379] env[62569]: return db.service_get_minimum_version(context, binaries) [ 456.084379] env[62569]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 456.084379] env[62569]: _check_db_access() [ 456.084379] env[62569]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 456.084379] env[62569]: stacktrace = ''.join(traceback.format_stack()) [ 456.084379] env[62569]: [ 456.085189] env[62569]: ERROR nova.db.main.api [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 456.085189] env[62569]: result = function(*args, **kwargs) [ 456.085189] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 456.085189] env[62569]: return func(*args, **kwargs) [ 456.085189] env[62569]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 456.085189] env[62569]: result = fn(*args, **kwargs) [ 456.085189] env[62569]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 456.085189] env[62569]: return f(*args, **kwargs) [ 456.085189] env[62569]: File "/opt/stack/nova/nova/objects/service.py", line 556, in _db_service_get_minimum_version [ 456.085189] env[62569]: return db.service_get_minimum_version(context, binaries) [ 456.085189] env[62569]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 456.085189] env[62569]: _check_db_access() [ 456.085189] env[62569]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 456.085189] env[62569]: stacktrace = ''.join(traceback.format_stack()) [ 456.085189] env[62569]: [ 456.085544] env[62569]: WARNING nova.objects.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] Failed to get minimum service version for cell 0649a166-844b-46c9-9b27-cb9b95b9bc15 [ 456.085699] env[62569]: WARNING nova.objects.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 456.086129] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] Acquiring lock "singleton_lock" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 456.086292] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] Acquired lock "singleton_lock" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 456.086541] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] Releasing lock "singleton_lock" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 456.086898] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] Full set of CONF: {{(pid=62569) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 456.087045] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ******************************************************************************** {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 456.087174] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] Configuration options gathered from: {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 456.087372] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 456.087602] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 456.087755] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ================================================================================ {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 456.087977] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] allow_resize_to_same_host = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.088161] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] arq_binding_timeout = 300 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.088291] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] backdoor_port = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.088417] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] backdoor_socket = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.088578] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] block_device_allocate_retries = 60 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.088795] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] block_device_allocate_retries_interval = 3 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.088980] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cert = self.pem {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.089159] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.089325] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] compute_monitors = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.089490] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] config_dir = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.089678] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] config_drive_format = iso9660 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.089886] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.090070] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] config_source = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.090243] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] console_host = devstack {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.090406] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] control_exchange = nova {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.090803] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cpu_allocation_ratio = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.090803] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] daemon = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.091619] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] debug = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.091619] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] default_access_ip_network_name = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.091619] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] default_availability_zone = nova {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.091619] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] default_ephemeral_format = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.091619] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] default_green_pool_size = 1000 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.091752] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.091909] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] default_schedule_zone = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.092069] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] disk_allocation_ratio = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.092233] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] enable_new_services = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.092406] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] enabled_apis = ['osapi_compute'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.092567] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] enabled_ssl_apis = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.092723] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] flat_injected = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.092882] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] force_config_drive = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.093050] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] force_raw_images = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.093223] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] graceful_shutdown_timeout = 5 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.093384] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] heal_instance_info_cache_interval = 60 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.093616] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] host = cpu-1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.093894] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.093986] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] initial_disk_allocation_ratio = 1.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.094167] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] initial_ram_allocation_ratio = 1.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.094388] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.094555] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] instance_build_timeout = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.094717] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] instance_delete_interval = 300 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.094883] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] instance_format = [instance: %(uuid)s] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.095056] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] instance_name_template = instance-%08x {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.095226] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] instance_usage_audit = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.095392] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] instance_usage_audit_period = month {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.095552] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.095715] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] instances_path = /opt/stack/data/nova/instances {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.095878] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] internal_service_availability_zone = internal {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.096042] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] key = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.096207] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] live_migration_retry_count = 30 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.096374] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] log_color = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.096537] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] log_config_append = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.096732] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.096889] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] log_dir = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.097060] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] log_file = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.097192] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] log_options = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.097355] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] log_rotate_interval = 1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.097522] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] log_rotate_interval_type = days {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.097686] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] log_rotation_type = none {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.097817] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.097943] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.098119] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.098284] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.098412] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.098571] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] long_rpc_timeout = 1800 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.098805] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] max_concurrent_builds = 10 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.098982] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] max_concurrent_live_migrations = 1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.099154] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] max_concurrent_snapshots = 5 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.099313] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] max_local_block_devices = 3 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.099473] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] max_logfile_count = 30 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.099666] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] max_logfile_size_mb = 200 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.099897] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] maximum_instance_delete_attempts = 5 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.100079] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] metadata_listen = 0.0.0.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.100250] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] metadata_listen_port = 8775 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.100419] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] metadata_workers = 2 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.100580] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] migrate_max_retries = -1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.100758] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] mkisofs_cmd = genisoimage {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.101033] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] my_block_storage_ip = 10.180.1.21 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.101179] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] my_ip = 10.180.1.21 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.101391] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] my_shared_fs_storage_ip = 10.180.1.21 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.101558] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] network_allocate_retries = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.101740] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.101909] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] osapi_compute_listen = 0.0.0.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.102083] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] osapi_compute_listen_port = 8774 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.102253] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] osapi_compute_unique_server_name_scope = {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.102422] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] osapi_compute_workers = 2 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.102584] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] password_length = 12 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.102742] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] periodic_enable = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.102902] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] periodic_fuzzy_delay = 60 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.103076] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] pointer_model = usbtablet {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.103508] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] preallocate_images = none {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.103689] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] publish_errors = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.103832] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] pybasedir = /opt/stack/nova {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.103987] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ram_allocation_ratio = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.104165] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] rate_limit_burst = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.104333] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] rate_limit_except_level = CRITICAL {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.104491] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] rate_limit_interval = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.104650] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] reboot_timeout = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.104811] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] reclaim_instance_interval = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.104965] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] record = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.105207] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] reimage_timeout_per_gb = 60 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.105393] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] report_interval = 120 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.105559] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] rescue_timeout = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.105721] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] reserved_host_cpus = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.105884] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] reserved_host_disk_mb = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.106054] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] reserved_host_memory_mb = 512 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.106312] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] reserved_huge_pages = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.106386] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] resize_confirm_window = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.106594] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] resize_fs_using_block_device = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.106718] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] resume_guests_state_on_host_boot = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.106891] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.107064] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] rpc_response_timeout = 60 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.107229] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] run_external_periodic_tasks = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.107396] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] running_deleted_instance_action = reap {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.107558] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] running_deleted_instance_poll_interval = 1800 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.107719] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] running_deleted_instance_timeout = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.107880] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] scheduler_instance_sync_interval = 120 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.108057] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] service_down_time = 720 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.108228] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] servicegroup_driver = db {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.108384] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] shell_completion = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.108543] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] shelved_offload_time = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.108745] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] shelved_poll_interval = 3600 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.108924] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] shutdown_timeout = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.109099] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] source_is_ipv6 = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.109262] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ssl_only = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.109505] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.109680] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] sync_power_state_interval = 600 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.109843] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] sync_power_state_pool_size = 1000 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.110015] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] syslog_log_facility = LOG_USER {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.110175] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] tempdir = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.110335] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] timeout_nbd = 10 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.110499] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] transport_url = **** {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.110660] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] update_resources_interval = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.110823] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] use_cow_images = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.110979] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] use_eventlog = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.111148] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] use_journal = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.111304] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] use_json = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.111460] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] use_rootwrap_daemon = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.111615] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] use_stderr = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.111770] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] use_syslog = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.111924] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vcpu_pin_set = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.112117] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vif_plugging_is_fatal = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.112389] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vif_plugging_timeout = 300 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.112670] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] virt_mkfs = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.112954] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] volume_usage_poll_interval = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.113230] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] watch_log_file = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.113431] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] web = /usr/share/spice-html5 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 456.113627] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.113800] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.113972] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.114162] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_concurrency.disable_process_locking = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.114456] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.114641] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.114810] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.114983] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.115171] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.115338] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.115518] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api.auth_strategy = keystone {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.115686] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api.compute_link_prefix = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.115863] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.116051] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api.dhcp_domain = novalocal {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.116225] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api.enable_instance_password = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.116388] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api.glance_link_prefix = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.116563] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.116754] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.116923] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api.instance_list_per_project_cells = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.117100] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api.list_records_by_skipping_down_cells = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.117266] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api.local_metadata_per_cell = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.117437] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api.max_limit = 1000 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.117604] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api.metadata_cache_expiration = 15 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.117778] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api.neutron_default_tenant_id = default {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.117953] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api.response_validation = warn {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.118134] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api.use_neutron_default_nets = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.118302] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.118467] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.118635] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.118844] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.119032] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api.vendordata_dynamic_targets = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.119200] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api.vendordata_jsonfile_path = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.119379] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.119570] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.backend = dogpile.cache.memcached {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.119757] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.backend_argument = **** {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.119936] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.config_prefix = cache.oslo {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.120128] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.dead_timeout = 60.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.120295] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.debug_cache_backend = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.120457] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.enable_retry_client = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.120618] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.enable_socket_keepalive = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.120788] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.enabled = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.120952] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.enforce_fips_mode = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.121129] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.expiration_time = 600 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.121294] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.hashclient_retry_attempts = 2 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.121459] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.hashclient_retry_delay = 1.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.121623] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.memcache_dead_retry = 300 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.121783] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.memcache_password = **** {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.121947] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.122122] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.122287] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.memcache_pool_maxsize = 10 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.122449] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.122611] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.memcache_sasl_enabled = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.122789] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.122958] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.memcache_socket_timeout = 1.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.123139] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.memcache_username = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.123306] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.proxies = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.123470] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.redis_db = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.123629] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.redis_password = **** {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.123796] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.redis_sentinel_service_name = mymaster {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.123972] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.124154] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.redis_server = localhost:6379 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.124320] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.redis_socket_timeout = 1.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.124477] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.redis_username = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.124637] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.retry_attempts = 2 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.124799] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.retry_delay = 0.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.124960] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.socket_keepalive_count = 1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.125131] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.socket_keepalive_idle = 1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.125292] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.socket_keepalive_interval = 1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.125445] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.tls_allowed_ciphers = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.125599] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.tls_cafile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.125753] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.tls_certfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.125912] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.tls_enabled = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.126082] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cache.tls_keyfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.126254] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cinder.auth_section = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.126425] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cinder.auth_type = password {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.126601] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cinder.cafile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.126780] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cinder.catalog_info = volumev3::publicURL {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.126941] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cinder.certfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.127116] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cinder.collect_timing = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.127279] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cinder.cross_az_attach = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.127442] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cinder.debug = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.127601] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cinder.endpoint_template = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.127764] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cinder.http_retries = 3 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.127928] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cinder.insecure = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.128096] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cinder.keyfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.128268] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cinder.os_region_name = RegionOne {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.128431] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cinder.split_loggers = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.128589] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cinder.timeout = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.128779] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.128944] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] compute.cpu_dedicated_set = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.129112] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] compute.cpu_shared_set = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.129279] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] compute.image_type_exclude_list = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.129441] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.129604] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] compute.max_concurrent_disk_ops = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.129787] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] compute.max_disk_devices_to_attach = -1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.129955] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.130135] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.130297] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] compute.resource_provider_association_refresh = 300 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.130457] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.130618] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] compute.shutdown_retry_interval = 10 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.130795] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.130974] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] conductor.workers = 2 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.131165] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] console.allowed_origins = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.131325] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] console.ssl_ciphers = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.131492] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] console.ssl_minimum_version = default {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.131657] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] consoleauth.enforce_session_timeout = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.131825] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] consoleauth.token_ttl = 600 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.131995] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cyborg.cafile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.132163] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cyborg.certfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.132325] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cyborg.collect_timing = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.132485] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cyborg.connect_retries = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.132641] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cyborg.connect_retry_delay = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.132796] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cyborg.endpoint_override = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.132959] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cyborg.insecure = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.133124] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cyborg.keyfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.133285] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cyborg.max_version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.133439] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cyborg.min_version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.133592] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cyborg.region_name = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.133745] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cyborg.retriable_status_codes = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.133900] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cyborg.service_name = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.134075] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cyborg.service_type = accelerator {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.134236] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cyborg.split_loggers = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.134390] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cyborg.status_code_retries = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.134545] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cyborg.status_code_retry_delay = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.134699] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cyborg.timeout = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.134880] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.135045] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] cyborg.version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.135226] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] database.backend = sqlalchemy {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.135392] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] database.connection = **** {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.135556] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] database.connection_debug = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.135721] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] database.connection_parameters = {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.135886] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] database.connection_recycle_time = 3600 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.136059] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] database.connection_trace = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.136225] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] database.db_inc_retry_interval = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.136420] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] database.db_max_retries = 20 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.136593] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] database.db_max_retry_interval = 10 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.137647] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] database.db_retry_interval = 1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.137647] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] database.max_overflow = 50 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.137647] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] database.max_pool_size = 5 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.137647] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] database.max_retries = 10 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.137647] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.137647] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] database.mysql_wsrep_sync_wait = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.137858] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] database.pool_timeout = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.137858] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] database.retry_interval = 10 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.137981] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] database.slave_connection = **** {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.138152] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] database.sqlite_synchronous = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.138312] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] database.use_db_reconnect = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.138487] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api_database.backend = sqlalchemy {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.138680] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api_database.connection = **** {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.138938] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api_database.connection_debug = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.139175] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api_database.connection_parameters = {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.139379] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api_database.connection_recycle_time = 3600 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.139555] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api_database.connection_trace = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.142910] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api_database.db_inc_retry_interval = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.142910] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api_database.db_max_retries = 20 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.142910] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api_database.db_max_retry_interval = 10 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.142910] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api_database.db_retry_interval = 1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.142910] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api_database.max_overflow = 50 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.142910] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api_database.max_pool_size = 5 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.142910] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api_database.max_retries = 10 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.143302] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.143302] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.143302] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api_database.pool_timeout = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.143302] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api_database.retry_interval = 10 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.143302] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api_database.slave_connection = **** {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.143302] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] api_database.sqlite_synchronous = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.143302] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] devices.enabled_mdev_types = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.143597] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.143597] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ephemeral_storage_encryption.default_format = luks {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.143597] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ephemeral_storage_encryption.enabled = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.143597] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.143597] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.api_servers = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.143597] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.cafile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.143597] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.certfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.143897] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.collect_timing = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.143897] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.connect_retries = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.143897] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.connect_retry_delay = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.143897] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.debug = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.143897] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.default_trusted_certificate_ids = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.144124] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.enable_certificate_validation = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.144124] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.enable_rbd_download = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.144286] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.endpoint_override = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.144452] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.insecure = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.144609] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.keyfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.144765] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.max_version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.144921] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.min_version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.145093] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.num_retries = 3 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.145264] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.rbd_ceph_conf = {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.145423] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.rbd_connect_timeout = 5 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.145588] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.rbd_pool = {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.145751] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.rbd_user = {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.145910] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.region_name = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.146077] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.retriable_status_codes = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.146237] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.service_name = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.146401] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.service_type = image {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.146580] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.split_loggers = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.146750] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.status_code_retries = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.146911] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.status_code_retry_delay = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.147077] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.timeout = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.147261] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.147426] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.verify_glance_signatures = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.147582] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] glance.version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.147746] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] guestfs.debug = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.147912] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] manila.auth_section = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.148086] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] manila.auth_type = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.148246] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] manila.cafile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.148400] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] manila.certfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.148562] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] manila.collect_timing = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.148744] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] manila.connect_retries = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.148914] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] manila.connect_retry_delay = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.149085] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] manila.endpoint_override = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.149248] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] manila.insecure = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.149402] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] manila.keyfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.149565] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] manila.max_version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.149746] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] manila.min_version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.149909] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] manila.region_name = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.150078] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] manila.retriable_status_codes = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.150236] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] manila.service_name = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.150404] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] manila.service_type = shared-file-system {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.150567] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] manila.share_apply_policy_timeout = 10 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.150753] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] manila.split_loggers = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.150929] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] manila.status_code_retries = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.151102] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] manila.status_code_retry_delay = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.151262] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] manila.timeout = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.151441] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] manila.valid_interfaces = ['internal', 'public'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.151598] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] manila.version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.151766] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] mks.enabled = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.152131] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.152322] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] image_cache.manager_interval = 2400 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.152490] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] image_cache.precache_concurrency = 1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.152661] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] image_cache.remove_unused_base_images = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.152831] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.152998] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.153188] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] image_cache.subdirectory_name = _base {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.153363] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.api_max_retries = 60 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.153529] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.api_retry_interval = 2 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.153688] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.auth_section = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.153848] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.auth_type = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.154013] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.cafile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.154179] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.certfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.154343] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.collect_timing = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.154503] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.conductor_group = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.154662] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.connect_retries = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.154823] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.connect_retry_delay = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.154977] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.endpoint_override = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.155154] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.insecure = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.155311] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.keyfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.155468] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.max_version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.155625] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.min_version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.155786] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.peer_list = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.155946] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.region_name = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.156120] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.retriable_status_codes = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.156283] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.serial_console_state_timeout = 10 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.156440] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.service_name = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.156624] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.service_type = baremetal {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.156794] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.shard = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.156961] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.split_loggers = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.157132] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.status_code_retries = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.157292] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.status_code_retry_delay = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.157450] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.timeout = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.157628] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.157787] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ironic.version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.157969] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.158154] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] key_manager.fixed_key = **** {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.158338] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.158501] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican.barbican_api_version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.158685] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican.barbican_endpoint = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.158874] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican.barbican_endpoint_type = public {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.159048] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican.barbican_region_name = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.159212] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican.cafile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.159373] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican.certfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.159537] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican.collect_timing = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.159730] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican.insecure = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.159933] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican.keyfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.160122] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican.number_of_retries = 60 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.160289] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican.retry_delay = 1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.160453] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican.send_service_user_token = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.160616] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican.split_loggers = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.160774] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican.timeout = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.160937] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican.verify_ssl = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.161106] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican.verify_ssl_path = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.161276] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican_service_user.auth_section = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.161436] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican_service_user.auth_type = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.161594] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican_service_user.cafile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.161748] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican_service_user.certfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.161911] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican_service_user.collect_timing = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.162080] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican_service_user.insecure = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.162313] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican_service_user.keyfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.162482] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican_service_user.split_loggers = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.162639] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] barbican_service_user.timeout = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.162818] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vault.approle_role_id = **** {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.162983] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vault.approle_secret_id = **** {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.163168] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vault.kv_mountpoint = secret {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.163327] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vault.kv_path = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.163489] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vault.kv_version = 2 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.163647] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vault.namespace = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.163804] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vault.root_token_id = **** {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.163962] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vault.ssl_ca_crt_file = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.164177] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vault.timeout = 60.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.164375] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vault.use_ssl = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.164552] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.164724] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] keystone.auth_section = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.164890] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] keystone.auth_type = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.165062] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] keystone.cafile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.165222] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] keystone.certfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.165384] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] keystone.collect_timing = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.165543] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] keystone.connect_retries = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.165701] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] keystone.connect_retry_delay = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.165861] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] keystone.endpoint_override = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.166031] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] keystone.insecure = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.166191] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] keystone.keyfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.166345] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] keystone.max_version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.166501] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] keystone.min_version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.166657] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] keystone.region_name = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.166815] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] keystone.retriable_status_codes = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.166971] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] keystone.service_name = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.167159] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] keystone.service_type = identity {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.167329] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] keystone.split_loggers = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.167489] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] keystone.status_code_retries = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.167644] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] keystone.status_code_retry_delay = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.167800] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] keystone.timeout = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.167978] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.168152] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] keystone.version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.168338] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.ceph_mount_options = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.168645] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.ceph_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.168852] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.connection_uri = {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.169031] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.cpu_mode = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.169205] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.cpu_model_extra_flags = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.169373] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.cpu_models = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.169541] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.cpu_power_governor_high = performance {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.169731] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.cpu_power_governor_low = powersave {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.169936] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.cpu_power_management = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.170127] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.170297] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.device_detach_attempts = 8 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.170462] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.device_detach_timeout = 20 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.170631] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.disk_cachemodes = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.170792] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.disk_prefix = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.170964] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.enabled_perf_events = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.171140] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.file_backed_memory = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.171309] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.gid_maps = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.171470] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.hw_disk_discard = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.171629] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.hw_machine_type = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.171799] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.images_rbd_ceph_conf = {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.171967] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.172143] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.172314] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.images_rbd_glance_store_name = {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.172483] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.images_rbd_pool = rbd {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.172655] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.images_type = default {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.172819] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.images_volume_group = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.172984] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.inject_key = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.173156] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.inject_partition = -2 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.173319] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.inject_password = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.173481] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.iscsi_iface = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.173642] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.iser_use_multipath = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.173806] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.live_migration_bandwidth = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.174027] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.174205] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.live_migration_downtime = 500 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.174371] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.174534] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.174697] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.live_migration_inbound_addr = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.174863] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.175033] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.live_migration_permit_post_copy = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.175199] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.live_migration_scheme = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.175368] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.live_migration_timeout_action = abort {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.175532] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.live_migration_tunnelled = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.175691] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.live_migration_uri = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.175855] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.live_migration_with_native_tls = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.176020] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.max_queues = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.176188] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.176411] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.176576] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.nfs_mount_options = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.176854] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.177040] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.177212] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.num_iser_scan_tries = 5 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.177372] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.num_memory_encrypted_guests = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.177534] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.177697] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.num_pcie_ports = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.177867] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.num_volume_scan_tries = 5 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.178043] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.pmem_namespaces = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.178210] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.quobyte_client_cfg = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.178487] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.178686] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.rbd_connect_timeout = 5 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.178864] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.179043] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.179209] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.rbd_secret_uuid = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.179368] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.rbd_user = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.179532] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.179745] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.remote_filesystem_transport = ssh {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.180020] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.rescue_image_id = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.180291] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.rescue_kernel_id = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.180486] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.rescue_ramdisk_id = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.180666] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.180833] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.rx_queue_size = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.181016] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.smbfs_mount_options = {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.181307] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.181483] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.snapshot_compression = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.181648] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.snapshot_image_format = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.181870] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.182053] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.sparse_logical_volumes = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.182223] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.swtpm_enabled = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.182395] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.swtpm_group = tss {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.182564] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.swtpm_user = tss {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.182732] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.sysinfo_serial = unique {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.182894] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.tb_cache_size = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.183062] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.tx_queue_size = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.183229] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.uid_maps = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.183394] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.use_virtio_for_bridges = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.183563] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.virt_type = kvm {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.183732] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.volume_clear = zero {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.183902] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.volume_clear_size = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.184079] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.volume_use_multipath = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.184243] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.vzstorage_cache_path = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.184412] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.184580] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.vzstorage_mount_group = qemu {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.184747] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.vzstorage_mount_opts = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.184972] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.185283] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.185466] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.vzstorage_mount_user = stack {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.185637] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.185811] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.auth_section = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.185988] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.auth_type = password {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.186166] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.cafile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.186329] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.certfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.186495] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.collect_timing = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.186656] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.connect_retries = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.186815] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.connect_retry_delay = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.186985] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.default_floating_pool = public {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.187158] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.endpoint_override = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.187323] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.extension_sync_interval = 600 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.187484] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.http_retries = 3 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.187643] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.insecure = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.187801] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.keyfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.187972] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.max_version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.188159] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.188321] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.min_version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.188488] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.ovs_bridge = br-int {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.188668] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.physnets = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.188842] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.region_name = RegionOne {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.189014] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.retriable_status_codes = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.189192] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.service_metadata_proxy = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.189350] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.service_name = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.189515] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.service_type = network {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.189701] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.split_loggers = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.189873] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.status_code_retries = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.190045] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.status_code_retry_delay = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.190210] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.timeout = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.190391] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.190551] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] neutron.version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.190723] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] notifications.bdms_in_notifications = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.190904] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] notifications.default_level = INFO {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.191089] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] notifications.notification_format = unversioned {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.191255] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] notifications.notify_on_state_change = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.191432] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.191603] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] pci.alias = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.191770] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] pci.device_spec = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.191936] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] pci.report_in_placement = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.192119] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.auth_section = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.192291] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.auth_type = password {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.192458] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.192617] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.cafile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.192774] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.certfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.192937] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.collect_timing = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.193106] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.connect_retries = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.193268] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.connect_retry_delay = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.193427] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.default_domain_id = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.193586] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.default_domain_name = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.193744] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.domain_id = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.193902] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.domain_name = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.194069] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.endpoint_override = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.194234] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.insecure = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.194391] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.keyfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.194545] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.max_version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.194699] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.min_version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.194867] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.password = **** {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.195039] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.project_domain_id = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.195210] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.project_domain_name = Default {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.195377] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.project_id = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.195548] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.project_name = service {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.195713] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.region_name = RegionOne {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.195875] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.retriable_status_codes = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.196040] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.service_name = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.196209] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.service_type = placement {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.196368] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.split_loggers = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.196526] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.status_code_retries = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.196684] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.status_code_retry_delay = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.196842] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.system_scope = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.196997] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.timeout = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.197167] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.trust_id = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.197322] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.user_domain_id = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.197486] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.user_domain_name = Default {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.197642] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.user_id = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.197813] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.username = nova {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.197991] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.198163] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] placement.version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.198339] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] quota.cores = 20 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.198507] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] quota.count_usage_from_placement = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.198718] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.198877] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] quota.injected_file_content_bytes = 10240 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.199057] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] quota.injected_file_path_length = 255 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.199338] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] quota.injected_files = 5 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.199517] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] quota.instances = 10 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.199702] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] quota.key_pairs = 100 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.199884] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] quota.metadata_items = 128 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.200064] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] quota.ram = 51200 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.200233] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] quota.recheck_quota = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.200401] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] quota.server_group_members = 10 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.200564] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] quota.server_groups = 10 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.200737] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.200904] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.201078] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] scheduler.image_metadata_prefilter = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.201241] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.201404] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] scheduler.max_attempts = 3 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.201566] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] scheduler.max_placement_results = 1000 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.201728] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.201890] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] scheduler.query_placement_for_image_type_support = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.202058] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.202233] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] scheduler.workers = 2 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.202407] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.202576] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.202754] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.202923] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.203098] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.203262] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.203426] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.203612] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.203777] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.host_subset_size = 1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.203941] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.204108] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.204272] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.204434] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.isolated_hosts = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.204592] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.isolated_images = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.204751] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.204910] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.205082] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.205248] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.pci_in_placement = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.205409] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.205583] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.205743] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.205903] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.206072] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.206234] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.206393] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.track_instance_changes = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.206567] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.206741] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] metrics.required = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.206910] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] metrics.weight_multiplier = 1.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.207084] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.207250] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] metrics.weight_setting = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.207573] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.207752] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] serial_console.enabled = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.207931] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] serial_console.port_range = 10000:20000 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.208110] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.208281] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.208447] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] serial_console.serialproxy_port = 6083 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.208611] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] service_user.auth_section = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.208808] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] service_user.auth_type = password {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.208974] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] service_user.cafile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.209144] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] service_user.certfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.209306] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] service_user.collect_timing = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.209464] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] service_user.insecure = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.209620] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] service_user.keyfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.209816] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] service_user.send_service_user_token = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.209987] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] service_user.split_loggers = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.210157] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] service_user.timeout = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.210324] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] spice.agent_enabled = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.210483] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] spice.enabled = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.210778] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.210971] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.211175] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] spice.html5proxy_port = 6082 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.211339] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] spice.image_compression = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.211496] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] spice.jpeg_compression = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.211655] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] spice.playback_compression = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.211817] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] spice.require_secure = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.211982] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] spice.server_listen = 127.0.0.1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.212162] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.212320] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] spice.streaming_mode = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.212474] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] spice.zlib_compression = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.212638] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] upgrade_levels.baseapi = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.212804] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] upgrade_levels.compute = auto {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.212962] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] upgrade_levels.conductor = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.213130] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] upgrade_levels.scheduler = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.213292] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vendordata_dynamic_auth.auth_section = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.213449] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vendordata_dynamic_auth.auth_type = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.213602] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vendordata_dynamic_auth.cafile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.213755] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vendordata_dynamic_auth.certfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.213918] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.214086] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vendordata_dynamic_auth.insecure = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.214243] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vendordata_dynamic_auth.keyfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.214402] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.214559] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vendordata_dynamic_auth.timeout = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.214730] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.api_retry_count = 10 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.214889] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.ca_file = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.215067] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.cache_prefix = devstack-image-cache {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.215233] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.cluster_name = testcl1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.215396] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.connection_pool_size = 10 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.215553] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.console_delay_seconds = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.215719] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.datastore_regex = ^datastore.* {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.215920] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.216101] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.host_password = **** {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.216266] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.host_port = 443 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.216433] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.host_username = administrator@vsphere.local {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.216599] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.insecure = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.216755] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.integration_bridge = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.216917] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.maximum_objects = 100 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.217085] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.pbm_default_policy = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.217250] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.pbm_enabled = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.217408] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.pbm_wsdl_location = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.217577] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.217735] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.serial_port_proxy_uri = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.217895] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.serial_port_service_uri = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.218074] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.task_poll_interval = 0.5 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.218245] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.use_linked_clone = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.218413] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.vnc_keymap = en-us {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.218576] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.vnc_port = 5900 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.218767] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vmware.vnc_port_total = 10000 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.218957] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vnc.auth_schemes = ['none'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.219146] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vnc.enabled = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.219437] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.219618] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.219808] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vnc.novncproxy_port = 6080 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.219989] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vnc.server_listen = 127.0.0.1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.220173] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.220334] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vnc.vencrypt_ca_certs = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.220490] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vnc.vencrypt_client_cert = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.220644] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vnc.vencrypt_client_key = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.220822] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.221010] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] workarounds.disable_deep_image_inspection = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.221192] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.221355] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.221515] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.221677] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] workarounds.disable_rootwrap = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.221839] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] workarounds.enable_numa_live_migration = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.221999] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.222171] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.222331] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.222491] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] workarounds.libvirt_disable_apic = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.222650] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.222813] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.222973] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.223144] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.223305] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.223463] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.223619] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.223776] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.223936] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.224110] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.224294] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.224462] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] wsgi.client_socket_timeout = 900 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.224628] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] wsgi.default_pool_size = 1000 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.224793] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] wsgi.keep_alive = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.224960] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] wsgi.max_header_line = 16384 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.225146] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] wsgi.secure_proxy_ssl_header = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.225309] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] wsgi.ssl_ca_file = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.225468] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] wsgi.ssl_cert_file = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.225627] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] wsgi.ssl_key_file = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.225789] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] wsgi.tcp_keepidle = 600 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.225969] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.226147] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] zvm.ca_file = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.226309] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] zvm.cloud_connector_url = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.226590] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.226764] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] zvm.reachable_timeout = 300 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.226945] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_policy.enforce_new_defaults = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.227368] env[62569]: WARNING oslo_config.cfg [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 456.227562] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_policy.enforce_scope = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.227743] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_policy.policy_default_rule = default {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.227930] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.228122] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_policy.policy_file = policy.yaml {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.228301] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.228465] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.228628] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.228816] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.228988] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.229170] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.229347] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.229526] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] profiler.connection_string = messaging:// {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.229723] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] profiler.enabled = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.229921] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] profiler.es_doc_type = notification {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.230104] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] profiler.es_scroll_size = 10000 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.230280] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] profiler.es_scroll_time = 2m {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.230447] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] profiler.filter_error_trace = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.230618] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] profiler.hmac_keys = **** {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.230788] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] profiler.sentinel_service_name = mymaster {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.230961] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] profiler.socket_timeout = 0.1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.231169] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] profiler.trace_requests = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.231340] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] profiler.trace_sqlalchemy = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.231525] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] profiler_jaeger.process_tags = {} {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.231688] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] profiler_jaeger.service_name_prefix = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.231852] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] profiler_otlp.service_name_prefix = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.232025] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] remote_debug.host = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.232186] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] remote_debug.port = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.232364] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.232526] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.232690] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.232852] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.233021] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.233249] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.233435] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.233600] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.233765] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.233938] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.234111] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.234282] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.234449] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.234619] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.234788] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.234957] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.235132] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.235306] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.235470] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.235632] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.235798] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.235970] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.236148] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.236317] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.236478] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.236640] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.236799] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.236961] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.237144] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.237309] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.ssl = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.237481] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.237649] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.237810] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.237979] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.238159] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.ssl_version = {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.238319] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.238507] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.238716] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_notifications.retry = -1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.238973] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.239182] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_messaging_notifications.transport_url = **** {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.239360] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_limit.auth_section = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.239525] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_limit.auth_type = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.239702] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_limit.cafile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.239877] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_limit.certfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.240056] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_limit.collect_timing = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.240223] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_limit.connect_retries = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.240386] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_limit.connect_retry_delay = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.240546] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_limit.endpoint_id = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.240706] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_limit.endpoint_override = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.240873] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_limit.insecure = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.241058] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_limit.keyfile = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.241231] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_limit.max_version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.241391] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_limit.min_version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.241550] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_limit.region_name = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.241712] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_limit.retriable_status_codes = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.241874] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_limit.service_name = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.242045] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_limit.service_type = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.242215] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_limit.split_loggers = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.242374] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_limit.status_code_retries = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.242533] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_limit.status_code_retry_delay = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.242693] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_limit.timeout = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.242855] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_limit.valid_interfaces = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.243022] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_limit.version = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.243193] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_reports.file_event_handler = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.243360] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.243519] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] oslo_reports.log_dir = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.243690] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.243852] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.244022] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.244186] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.244347] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.244504] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.244673] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.244833] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vif_plug_ovs_privileged.group = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.244990] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.245169] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.245333] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.245488] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] vif_plug_ovs_privileged.user = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.245657] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] os_vif_linux_bridge.flat_interface = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.245837] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.246014] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.246195] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.246364] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.246535] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.246701] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.246864] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.247056] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.247229] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] os_vif_ovs.isolate_vif = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.247397] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.247563] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.247732] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.247901] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] os_vif_ovs.ovsdb_interface = native {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.248075] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] os_vif_ovs.per_port_bridge = False {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.248247] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] privsep_osbrick.capabilities = [21] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.248406] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] privsep_osbrick.group = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.248562] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] privsep_osbrick.helper_command = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.248746] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.248921] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.249092] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] privsep_osbrick.user = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.249266] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.249424] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] nova_sys_admin.group = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.249579] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] nova_sys_admin.helper_command = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.249759] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.249931] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.250101] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] nova_sys_admin.user = None {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 456.250233] env[62569]: DEBUG oslo_service.service [None req-e97c2f2e-f224-49c3-8b90-a0ed074d6e90 None None] ******************************************************************************** {{(pid=62569) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 456.250722] env[62569]: INFO nova.service [-] Starting compute node (version 30.1.0) [ 456.754027] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Getting list of instances from cluster (obj){ [ 456.754027] env[62569]: value = "domain-c8" [ 456.754027] env[62569]: _type = "ClusterComputeResource" [ 456.754027] env[62569]: } {{(pid=62569) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 456.755265] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-423c06a5-6095-4ee3-a984-818bd259a4fe {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 456.764360] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Got total of 0 instances {{(pid=62569) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 456.764921] env[62569]: WARNING nova.virt.vmwareapi.driver [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 456.765360] env[62569]: INFO nova.virt.node [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Generated node identity fa06556a-5785-4014-b8bd-bc240a0cf716 [ 456.765586] env[62569]: INFO nova.virt.node [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Wrote node identity fa06556a-5785-4014-b8bd-bc240a0cf716 to /opt/stack/data/n-cpu-1/compute_id [ 457.268914] env[62569]: WARNING nova.compute.manager [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Compute nodes ['fa06556a-5785-4014-b8bd-bc240a0cf716'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 458.275521] env[62569]: INFO nova.compute.manager [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 459.281688] env[62569]: WARNING nova.compute.manager [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 459.282085] env[62569]: DEBUG oslo_concurrency.lockutils [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 459.282201] env[62569]: DEBUG oslo_concurrency.lockutils [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 459.282355] env[62569]: DEBUG oslo_concurrency.lockutils [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 459.282510] env[62569]: DEBUG nova.compute.resource_tracker [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62569) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 459.283478] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99f59b96-fbf0-4899-a087-013848451ea4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 459.291927] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f69dc8d-2b31-46bb-a1f7-a7a6d64a9481 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 459.306804] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04a7e023-7eb3-4b1f-80aa-a3392d6f2f70 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 459.312991] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cb08d89-ed83-48bc-95dd-cf047bfd0dc7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 459.341323] env[62569]: DEBUG nova.compute.resource_tracker [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181555MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=62569) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 459.341495] env[62569]: DEBUG oslo_concurrency.lockutils [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 459.341659] env[62569]: DEBUG oslo_concurrency.lockutils [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 459.844760] env[62569]: WARNING nova.compute.resource_tracker [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] No compute node record for cpu-1:fa06556a-5785-4014-b8bd-bc240a0cf716: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host fa06556a-5785-4014-b8bd-bc240a0cf716 could not be found. [ 460.352130] env[62569]: INFO nova.compute.resource_tracker [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: fa06556a-5785-4014-b8bd-bc240a0cf716 [ 461.860257] env[62569]: DEBUG nova.compute.resource_tracker [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62569) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 461.860644] env[62569]: DEBUG nova.compute.resource_tracker [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62569) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 462.020394] env[62569]: INFO nova.scheduler.client.report [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] [req-46c2ea07-24f8-4e04-8af0-d46abca90b19] Created resource provider record via placement API for resource provider with UUID fa06556a-5785-4014-b8bd-bc240a0cf716 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 462.037743] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c394f85-d352-4334-a4e1-534dd50f9ebf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 462.045197] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b661edd-580f-48b4-b196-8cfba10c27ef {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 462.074488] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c99acf31-1091-4f2c-9965-06e01c4b2b4d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 462.081143] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb13996e-6cef-47da-acc2-aba52e53dadf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 462.093617] env[62569]: DEBUG nova.compute.provider_tree [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 462.627559] env[62569]: DEBUG nova.scheduler.client.report [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Updated inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 462.627784] env[62569]: DEBUG nova.compute.provider_tree [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Updating resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 generation from 0 to 1 during operation: update_inventory {{(pid=62569) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 462.627945] env[62569]: DEBUG nova.compute.provider_tree [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 462.680408] env[62569]: DEBUG nova.compute.provider_tree [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Updating resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 generation from 1 to 2 during operation: update_traits {{(pid=62569) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 463.185619] env[62569]: DEBUG nova.compute.resource_tracker [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62569) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 463.185978] env[62569]: DEBUG oslo_concurrency.lockutils [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.844s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 463.185978] env[62569]: DEBUG nova.service [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Creating RPC server for service compute {{(pid=62569) start /opt/stack/nova/nova/service.py:186}} [ 463.199756] env[62569]: DEBUG nova.service [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] Join ServiceGroup membership for this service compute {{(pid=62569) start /opt/stack/nova/nova/service.py:203}} [ 463.199946] env[62569]: DEBUG nova.servicegroup.drivers.db [None req-79901844-11b6-45ee-b332-5f88f8c246a8 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62569) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 495.664474] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Acquiring lock "a2d5b36d-357f-42b0-8538-77c5ef915d61" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 495.664842] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Lock "a2d5b36d-357f-42b0-8538-77c5ef915d61" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 496.169238] env[62569]: DEBUG nova.compute.manager [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 496.716045] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 496.716383] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 496.718469] env[62569]: INFO nova.compute.claims [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 497.772018] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76953f3a-a45b-4c8d-960b-912f32fe744c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.779243] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bff6579b-f1b9-44d9-8efa-1b62d10dedb7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.816283] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aff82036-8b28-45ae-bf54-e72caa0c5fc0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.824680] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa73ed17-69d7-4a03-9a8b-df267243b9a9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.840771] env[62569]: DEBUG nova.compute.provider_tree [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 498.345018] env[62569]: DEBUG nova.scheduler.client.report [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 498.853670] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.137s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 498.854304] env[62569]: DEBUG nova.compute.manager [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 499.362690] env[62569]: DEBUG nova.compute.utils [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 499.363988] env[62569]: DEBUG nova.compute.manager [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 499.364867] env[62569]: DEBUG nova.network.neutron [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 499.872663] env[62569]: DEBUG nova.compute.manager [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 499.916290] env[62569]: DEBUG nova.policy [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6afea4ca95ea458aa26ad4ef74921072', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '73cf75e06b424da98d1bad15070ea90b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 500.198973] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Acquiring lock "99412ace-3825-4c62-ad54-e59b83218387" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 500.199471] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Lock "99412ace-3825-4c62-ad54-e59b83218387" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 500.336010] env[62569]: DEBUG nova.network.neutron [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Successfully created port: 6abfdbf1-8e5c-499b-98d7-789b91c9ee3a {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 500.464488] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Acquiring lock "17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 500.464736] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Lock "17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 500.702801] env[62569]: DEBUG nova.compute.manager [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 500.882506] env[62569]: DEBUG nova.compute.manager [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 500.918436] env[62569]: DEBUG nova.virt.hardware [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 500.918831] env[62569]: DEBUG nova.virt.hardware [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 500.918905] env[62569]: DEBUG nova.virt.hardware [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 500.920015] env[62569]: DEBUG nova.virt.hardware [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 500.920015] env[62569]: DEBUG nova.virt.hardware [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 500.920015] env[62569]: DEBUG nova.virt.hardware [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 500.920015] env[62569]: DEBUG nova.virt.hardware [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 500.920015] env[62569]: DEBUG nova.virt.hardware [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 500.920820] env[62569]: DEBUG nova.virt.hardware [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 500.921741] env[62569]: DEBUG nova.virt.hardware [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 500.921999] env[62569]: DEBUG nova.virt.hardware [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 500.923024] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b8010a-31ea-46e9-914d-73dead930f36 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.931759] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5f3aba8-220d-436c-a073-b250eaa7dd4a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.949919] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c89b0bb-32a9-48fc-8746-3bbc3bb550b8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.968120] env[62569]: DEBUG nova.compute.manager [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 501.054322] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Acquiring lock "3a9b16bb-9d58-4f0a-8fab-16f2486a6702" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 501.054583] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Lock "3a9b16bb-9d58-4f0a-8fab-16f2486a6702" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 501.247099] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 501.247505] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.004s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 501.249461] env[62569]: INFO nova.compute.claims [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 501.492267] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 501.557422] env[62569]: DEBUG nova.compute.manager [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 501.812246] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Acquiring lock "611d4cfa-a888-47bc-8857-f061046a77ed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 501.813542] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Lock "611d4cfa-a888-47bc-8857-f061046a77ed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 502.012253] env[62569]: ERROR nova.compute.manager [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6abfdbf1-8e5c-499b-98d7-789b91c9ee3a, please check neutron logs for more information. [ 502.012253] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 502.012253] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 502.012253] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 502.012253] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 502.012253] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 502.012253] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 502.012253] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 502.012253] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 502.012253] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 502.012253] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 502.012253] env[62569]: ERROR nova.compute.manager raise self.value [ 502.012253] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 502.012253] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 502.012253] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 502.012253] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 502.012950] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 502.012950] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 502.012950] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6abfdbf1-8e5c-499b-98d7-789b91c9ee3a, please check neutron logs for more information. [ 502.012950] env[62569]: ERROR nova.compute.manager [ 502.013676] env[62569]: Traceback (most recent call last): [ 502.017069] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 502.017069] env[62569]: listener.cb(fileno) [ 502.017069] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 502.017069] env[62569]: result = function(*args, **kwargs) [ 502.017069] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 502.017069] env[62569]: return func(*args, **kwargs) [ 502.017069] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 502.017069] env[62569]: raise e [ 502.017069] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 502.017069] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 502.017069] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 502.017069] env[62569]: created_port_ids = self._update_ports_for_instance( [ 502.017069] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 502.017069] env[62569]: with excutils.save_and_reraise_exception(): [ 502.017069] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 502.017069] env[62569]: self.force_reraise() [ 502.017069] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 502.017069] env[62569]: raise self.value [ 502.017069] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 502.017069] env[62569]: updated_port = self._update_port( [ 502.017069] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 502.017069] env[62569]: _ensure_no_port_binding_failure(port) [ 502.017069] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 502.017069] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 502.017069] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 6abfdbf1-8e5c-499b-98d7-789b91c9ee3a, please check neutron logs for more information. [ 502.017069] env[62569]: Removing descriptor: 14 [ 502.017883] env[62569]: ERROR nova.compute.manager [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6abfdbf1-8e5c-499b-98d7-789b91c9ee3a, please check neutron logs for more information. [ 502.017883] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Traceback (most recent call last): [ 502.017883] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 502.017883] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] yield resources [ 502.017883] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 502.017883] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] self.driver.spawn(context, instance, image_meta, [ 502.017883] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 502.017883] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] self._vmops.spawn(context, instance, image_meta, injected_files, [ 502.017883] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 502.017883] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] vm_ref = self.build_virtual_machine(instance, [ 502.017883] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 502.018266] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] vif_infos = vmwarevif.get_vif_info(self._session, [ 502.018266] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 502.018266] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] for vif in network_info: [ 502.018266] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 502.018266] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] return self._sync_wrapper(fn, *args, **kwargs) [ 502.018266] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 502.018266] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] self.wait() [ 502.018266] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 502.018266] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] self[:] = self._gt.wait() [ 502.018266] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 502.018266] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] return self._exit_event.wait() [ 502.018266] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 502.018266] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] result = hub.switch() [ 502.018832] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 502.018832] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] return self.greenlet.switch() [ 502.018832] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 502.018832] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] result = function(*args, **kwargs) [ 502.018832] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 502.018832] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] return func(*args, **kwargs) [ 502.018832] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 502.018832] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] raise e [ 502.018832] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 502.018832] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] nwinfo = self.network_api.allocate_for_instance( [ 502.018832] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 502.018832] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] created_port_ids = self._update_ports_for_instance( [ 502.018832] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 502.019237] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] with excutils.save_and_reraise_exception(): [ 502.019237] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 502.019237] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] self.force_reraise() [ 502.019237] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 502.019237] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] raise self.value [ 502.019237] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 502.019237] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] updated_port = self._update_port( [ 502.019237] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 502.019237] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] _ensure_no_port_binding_failure(port) [ 502.019237] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 502.019237] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] raise exception.PortBindingFailed(port_id=port['id']) [ 502.019237] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] nova.exception.PortBindingFailed: Binding failed for port 6abfdbf1-8e5c-499b-98d7-789b91c9ee3a, please check neutron logs for more information. [ 502.019237] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] [ 502.019622] env[62569]: INFO nova.compute.manager [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Terminating instance [ 502.091889] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 502.317158] env[62569]: DEBUG nova.compute.manager [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 502.373519] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc2461fc-e6ca-48ea-b5a7-eb9994729a59 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.382196] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8a17d0d-7abb-4461-8f17-dd72d8c77f52 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.419807] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05cfa136-0760-4afa-b1ae-f690f77928a0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.424980] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dbd2f1c-5dab-4c14-a939-235fa50dbc95 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.440713] env[62569]: DEBUG nova.compute.provider_tree [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 502.531779] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Acquiring lock "refresh_cache-a2d5b36d-357f-42b0-8538-77c5ef915d61" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 502.531779] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Acquired lock "refresh_cache-a2d5b36d-357f-42b0-8538-77c5ef915d61" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 502.531779] env[62569]: DEBUG nova.network.neutron [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 502.601685] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Acquiring lock "ff90249f-ae4b-4aeb-91ed-d8a41a942a5d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 502.601790] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Lock "ff90249f-ae4b-4aeb-91ed-d8a41a942a5d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 502.860688] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 502.928097] env[62569]: DEBUG nova.compute.manager [req-6ef6c1dd-8cbc-44aa-b392-9f0f2a1d89a6 req-8707ac39-9cf7-45f7-875a-bb30cf4507d8 service nova] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Received event network-changed-6abfdbf1-8e5c-499b-98d7-789b91c9ee3a {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 502.929028] env[62569]: DEBUG nova.compute.manager [req-6ef6c1dd-8cbc-44aa-b392-9f0f2a1d89a6 req-8707ac39-9cf7-45f7-875a-bb30cf4507d8 service nova] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Refreshing instance network info cache due to event network-changed-6abfdbf1-8e5c-499b-98d7-789b91c9ee3a. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 502.929028] env[62569]: DEBUG oslo_concurrency.lockutils [req-6ef6c1dd-8cbc-44aa-b392-9f0f2a1d89a6 req-8707ac39-9cf7-45f7-875a-bb30cf4507d8 service nova] Acquiring lock "refresh_cache-a2d5b36d-357f-42b0-8538-77c5ef915d61" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 502.943709] env[62569]: DEBUG nova.scheduler.client.report [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 503.075893] env[62569]: DEBUG nova.network.neutron [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 503.103872] env[62569]: DEBUG nova.compute.manager [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 503.171724] env[62569]: DEBUG nova.network.neutron [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 503.450371] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.203s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 503.451024] env[62569]: DEBUG nova.compute.manager [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 503.455908] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.964s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 503.457652] env[62569]: INFO nova.compute.claims [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 503.635536] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 503.677408] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Releasing lock "refresh_cache-a2d5b36d-357f-42b0-8538-77c5ef915d61" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 503.677717] env[62569]: DEBUG nova.compute.manager [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 503.677924] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 503.678286] env[62569]: DEBUG oslo_concurrency.lockutils [req-6ef6c1dd-8cbc-44aa-b392-9f0f2a1d89a6 req-8707ac39-9cf7-45f7-875a-bb30cf4507d8 service nova] Acquired lock "refresh_cache-a2d5b36d-357f-42b0-8538-77c5ef915d61" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 503.678541] env[62569]: DEBUG nova.network.neutron [req-6ef6c1dd-8cbc-44aa-b392-9f0f2a1d89a6 req-8707ac39-9cf7-45f7-875a-bb30cf4507d8 service nova] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Refreshing network info cache for port 6abfdbf1-8e5c-499b-98d7-789b91c9ee3a {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 503.682021] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-69450842-bada-4bc1-993a-687699a92c03 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.691288] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eadfbc27-a5e4-4e89-832e-c3374da720ed {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.715524] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a2d5b36d-357f-42b0-8538-77c5ef915d61 could not be found. [ 503.715524] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 503.716157] env[62569]: INFO nova.compute.manager [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Took 0.04 seconds to destroy the instance on the hypervisor. [ 503.716157] env[62569]: DEBUG oslo.service.loopingcall [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 503.716308] env[62569]: DEBUG nova.compute.manager [-] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 503.716394] env[62569]: DEBUG nova.network.neutron [-] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 503.751745] env[62569]: DEBUG nova.network.neutron [-] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 503.966377] env[62569]: DEBUG nova.compute.utils [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 503.972896] env[62569]: DEBUG nova.compute.manager [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 503.972896] env[62569]: DEBUG nova.network.neutron [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 504.048958] env[62569]: DEBUG nova.policy [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9eb4ef48bcde4bd780a7721f7e8fc230', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '705b55e8b5714144b451ba99f59fada7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 504.057223] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Acquiring lock "832ec210-bb14-4a52-9634-50b7709a9976" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 504.057519] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Lock "832ec210-bb14-4a52-9634-50b7709a9976" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 504.232785] env[62569]: DEBUG nova.network.neutron [req-6ef6c1dd-8cbc-44aa-b392-9f0f2a1d89a6 req-8707ac39-9cf7-45f7-875a-bb30cf4507d8 service nova] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 504.255751] env[62569]: DEBUG nova.network.neutron [-] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 504.393097] env[62569]: DEBUG nova.network.neutron [req-6ef6c1dd-8cbc-44aa-b392-9f0f2a1d89a6 req-8707ac39-9cf7-45f7-875a-bb30cf4507d8 service nova] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 504.475914] env[62569]: DEBUG nova.compute.manager [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 504.561793] env[62569]: DEBUG nova.compute.manager [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 504.634304] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5bfe201-ee62-4cb9-b03d-072264789d9f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.647080] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a195bd51-c4ed-4233-9851-173181585e27 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.656155] env[62569]: DEBUG nova.network.neutron [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Successfully created port: 8d597f38-b911-472d-8ad8-4a66203f2e55 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 504.688087] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59da604e-9315-421b-830d-55d05d4fbdf4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.695965] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01767808-dc11-45ee-9304-bceb8a493b5b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.710942] env[62569]: DEBUG nova.compute.provider_tree [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 504.758519] env[62569]: INFO nova.compute.manager [-] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Took 1.04 seconds to deallocate network for instance. [ 504.761147] env[62569]: DEBUG nova.compute.claims [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 504.761426] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 504.898401] env[62569]: DEBUG oslo_concurrency.lockutils [req-6ef6c1dd-8cbc-44aa-b392-9f0f2a1d89a6 req-8707ac39-9cf7-45f7-875a-bb30cf4507d8 service nova] Releasing lock "refresh_cache-a2d5b36d-357f-42b0-8538-77c5ef915d61" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 505.087485] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 505.214576] env[62569]: DEBUG nova.scheduler.client.report [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 505.492746] env[62569]: DEBUG nova.compute.manager [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 505.526254] env[62569]: DEBUG nova.virt.hardware [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 505.526254] env[62569]: DEBUG nova.virt.hardware [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 505.526254] env[62569]: DEBUG nova.virt.hardware [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 505.526475] env[62569]: DEBUG nova.virt.hardware [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 505.526475] env[62569]: DEBUG nova.virt.hardware [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 505.526475] env[62569]: DEBUG nova.virt.hardware [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 505.526943] env[62569]: DEBUG nova.virt.hardware [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 505.527425] env[62569]: DEBUG nova.virt.hardware [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 505.527777] env[62569]: DEBUG nova.virt.hardware [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 505.528084] env[62569]: DEBUG nova.virt.hardware [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 505.528400] env[62569]: DEBUG nova.virt.hardware [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 505.531032] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e914888-5a6a-4920-982c-92bbd5daa09a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.538566] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee313fc7-37f4-4c4c-b502-e1f493840a0b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.723812] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.267s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 505.727488] env[62569]: DEBUG nova.compute.manager [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 505.729991] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.638s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 505.732700] env[62569]: INFO nova.compute.claims [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 505.800522] env[62569]: ERROR nova.compute.manager [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8d597f38-b911-472d-8ad8-4a66203f2e55, please check neutron logs for more information. [ 505.800522] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 505.800522] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 505.800522] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 505.800522] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 505.800522] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 505.800522] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 505.800522] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 505.800522] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 505.800522] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 505.800522] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 505.800522] env[62569]: ERROR nova.compute.manager raise self.value [ 505.800522] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 505.800522] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 505.800522] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 505.800522] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 505.801074] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 505.801074] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 505.801074] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8d597f38-b911-472d-8ad8-4a66203f2e55, please check neutron logs for more information. [ 505.801074] env[62569]: ERROR nova.compute.manager [ 505.801074] env[62569]: Traceback (most recent call last): [ 505.801229] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 505.801229] env[62569]: listener.cb(fileno) [ 505.801229] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 505.801229] env[62569]: result = function(*args, **kwargs) [ 505.801229] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 505.801229] env[62569]: return func(*args, **kwargs) [ 505.801229] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 505.801229] env[62569]: raise e [ 505.801229] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 505.801229] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 505.801229] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 505.801229] env[62569]: created_port_ids = self._update_ports_for_instance( [ 505.801229] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 505.801229] env[62569]: with excutils.save_and_reraise_exception(): [ 505.801229] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 505.801229] env[62569]: self.force_reraise() [ 505.801229] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 505.801229] env[62569]: raise self.value [ 505.801229] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 505.801229] env[62569]: updated_port = self._update_port( [ 505.801229] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 505.801229] env[62569]: _ensure_no_port_binding_failure(port) [ 505.801229] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 505.801229] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 505.801229] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 8d597f38-b911-472d-8ad8-4a66203f2e55, please check neutron logs for more information. [ 505.801229] env[62569]: Removing descriptor: 14 [ 505.802754] env[62569]: ERROR nova.compute.manager [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8d597f38-b911-472d-8ad8-4a66203f2e55, please check neutron logs for more information. [ 505.802754] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] Traceback (most recent call last): [ 505.802754] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 505.802754] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] yield resources [ 505.802754] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 505.802754] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] self.driver.spawn(context, instance, image_meta, [ 505.802754] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 505.802754] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] self._vmops.spawn(context, instance, image_meta, injected_files, [ 505.802754] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 505.802754] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] vm_ref = self.build_virtual_machine(instance, [ 505.802754] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 505.803141] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] vif_infos = vmwarevif.get_vif_info(self._session, [ 505.803141] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 505.803141] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] for vif in network_info: [ 505.803141] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 505.803141] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] return self._sync_wrapper(fn, *args, **kwargs) [ 505.803141] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 505.803141] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] self.wait() [ 505.803141] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 505.803141] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] self[:] = self._gt.wait() [ 505.803141] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 505.803141] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] return self._exit_event.wait() [ 505.803141] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 505.803141] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] result = hub.switch() [ 505.804267] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 505.804267] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] return self.greenlet.switch() [ 505.804267] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 505.804267] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] result = function(*args, **kwargs) [ 505.804267] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 505.804267] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] return func(*args, **kwargs) [ 505.804267] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 505.804267] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] raise e [ 505.804267] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 505.804267] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] nwinfo = self.network_api.allocate_for_instance( [ 505.804267] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 505.804267] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] created_port_ids = self._update_ports_for_instance( [ 505.804267] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 505.804637] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] with excutils.save_and_reraise_exception(): [ 505.804637] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 505.804637] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] self.force_reraise() [ 505.804637] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 505.804637] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] raise self.value [ 505.804637] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 505.804637] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] updated_port = self._update_port( [ 505.804637] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 505.804637] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] _ensure_no_port_binding_failure(port) [ 505.804637] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 505.804637] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] raise exception.PortBindingFailed(port_id=port['id']) [ 505.804637] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] nova.exception.PortBindingFailed: Binding failed for port 8d597f38-b911-472d-8ad8-4a66203f2e55, please check neutron logs for more information. [ 505.804637] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] [ 505.805220] env[62569]: INFO nova.compute.manager [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Terminating instance [ 505.810427] env[62569]: DEBUG nova.compute.manager [req-bd34a71f-bfcc-41e9-829d-aeca2719ec6e req-97e38eae-3d50-4456-bb8e-80eb810c6529 service nova] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Received event network-vif-deleted-6abfdbf1-8e5c-499b-98d7-789b91c9ee3a {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 506.238339] env[62569]: DEBUG nova.compute.utils [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 506.242158] env[62569]: DEBUG nova.compute.manager [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Not allocating networking since 'none' was specified. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 506.311039] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Acquiring lock "refresh_cache-99412ace-3825-4c62-ad54-e59b83218387" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 506.311241] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Acquired lock "refresh_cache-99412ace-3825-4c62-ad54-e59b83218387" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 506.311445] env[62569]: DEBUG nova.network.neutron [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 506.743960] env[62569]: DEBUG nova.compute.manager [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 506.840861] env[62569]: DEBUG nova.network.neutron [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 506.888260] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a89f1157-48b5-4597-9a3b-79fa2ddc3fb3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.897638] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31ddb376-b366-4401-9d06-b1ae65da7c59 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.932543] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4459c11d-52bc-4fe8-9635-706c03defcaf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.940927] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90046d11-4c6b-49ce-815f-36b074cb4bc5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.955989] env[62569]: DEBUG nova.compute.provider_tree [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 506.986085] env[62569]: DEBUG nova.network.neutron [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 507.461509] env[62569]: DEBUG nova.scheduler.client.report [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 507.492332] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Releasing lock "refresh_cache-99412ace-3825-4c62-ad54-e59b83218387" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 507.492332] env[62569]: DEBUG nova.compute.manager [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 507.492332] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 507.492332] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8e45d00a-0a09-4209-81cc-5d21eb8eb041 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.501311] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cefa75c8-b746-4afc-8011-39167684d3dc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.528705] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 99412ace-3825-4c62-ad54-e59b83218387 could not be found. [ 507.529317] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 507.530024] env[62569]: INFO nova.compute.manager [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Took 0.04 seconds to destroy the instance on the hypervisor. [ 507.530260] env[62569]: DEBUG oslo.service.loopingcall [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 507.531058] env[62569]: DEBUG nova.compute.manager [-] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 507.531058] env[62569]: DEBUG nova.network.neutron [-] [instance: 99412ace-3825-4c62-ad54-e59b83218387] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 507.572727] env[62569]: DEBUG nova.network.neutron [-] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 507.763600] env[62569]: DEBUG nova.compute.manager [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 507.809550] env[62569]: DEBUG nova.virt.hardware [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 507.809784] env[62569]: DEBUG nova.virt.hardware [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 507.809947] env[62569]: DEBUG nova.virt.hardware [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 507.810146] env[62569]: DEBUG nova.virt.hardware [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 507.810293] env[62569]: DEBUG nova.virt.hardware [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 507.810443] env[62569]: DEBUG nova.virt.hardware [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 507.810649] env[62569]: DEBUG nova.virt.hardware [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 507.810806] env[62569]: DEBUG nova.virt.hardware [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 507.810968] env[62569]: DEBUG nova.virt.hardware [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 507.811151] env[62569]: DEBUG nova.virt.hardware [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 507.811345] env[62569]: DEBUG nova.virt.hardware [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 507.812255] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da857c1-06a2-47fe-8390-9019b7f7f0b2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.821714] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a762401-b4eb-4595-9664-5271363d642b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.835628] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Instance VIF info [] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 507.847087] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 507.847583] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2d73442e-0767-4260-a7b4-a3c1e50fa991 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.858841] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Created folder: OpenStack in parent group-v4. [ 507.859089] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Creating folder: Project (78900030f25546f697a4224e07890c9c). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 507.859336] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5449b407-0500-4c28-a26b-462cb4cd4915 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.869526] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Created folder: Project (78900030f25546f697a4224e07890c9c) in parent group-v269330. [ 507.869600] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Creating folder: Instances. Parent ref: group-v269331. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 507.869800] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-580ec984-3b54-4810-bb99-0306d0c4ba73 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.879469] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Created folder: Instances in parent group-v269331. [ 507.879773] env[62569]: DEBUG oslo.service.loopingcall [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 507.879942] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 507.880086] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b71e848a-7c61-4ecb-ab9f-7448ee3afb5d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.900495] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 507.900495] env[62569]: value = "task-1249789" [ 507.900495] env[62569]: _type = "Task" [ 507.900495] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 507.911713] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249789, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 507.965866] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.236s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 507.966519] env[62569]: DEBUG nova.compute.manager [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 507.969456] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.109s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 507.970747] env[62569]: INFO nova.compute.claims [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 508.076416] env[62569]: DEBUG nova.network.neutron [-] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 508.159851] env[62569]: DEBUG nova.compute.manager [req-17054613-6a5b-466f-85a4-35769530fe9c req-166526cf-1980-41a0-93f5-99858e878b48 service nova] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Received event network-changed-8d597f38-b911-472d-8ad8-4a66203f2e55 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 508.160191] env[62569]: DEBUG nova.compute.manager [req-17054613-6a5b-466f-85a4-35769530fe9c req-166526cf-1980-41a0-93f5-99858e878b48 service nova] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Refreshing instance network info cache due to event network-changed-8d597f38-b911-472d-8ad8-4a66203f2e55. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 508.160363] env[62569]: DEBUG oslo_concurrency.lockutils [req-17054613-6a5b-466f-85a4-35769530fe9c req-166526cf-1980-41a0-93f5-99858e878b48 service nova] Acquiring lock "refresh_cache-99412ace-3825-4c62-ad54-e59b83218387" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 508.160530] env[62569]: DEBUG oslo_concurrency.lockutils [req-17054613-6a5b-466f-85a4-35769530fe9c req-166526cf-1980-41a0-93f5-99858e878b48 service nova] Acquired lock "refresh_cache-99412ace-3825-4c62-ad54-e59b83218387" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 508.160657] env[62569]: DEBUG nova.network.neutron [req-17054613-6a5b-466f-85a4-35769530fe9c req-166526cf-1980-41a0-93f5-99858e878b48 service nova] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Refreshing network info cache for port 8d597f38-b911-472d-8ad8-4a66203f2e55 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 508.415833] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249789, 'name': CreateVM_Task, 'duration_secs': 0.321648} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 508.416123] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 508.417229] env[62569]: DEBUG oslo_vmware.service [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-790b8343-3473-4108-9a3c-8aad318a1421 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.424955] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 508.424955] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 508.424955] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 508.425317] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db03efdf-7d10-4caf-ad4b-808369243480 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.431199] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Waiting for the task: (returnval){ [ 508.431199] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f379df-f1f1-9760-a90b-d4ab0936c023" [ 508.431199] env[62569]: _type = "Task" [ 508.431199] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 508.438664] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f379df-f1f1-9760-a90b-d4ab0936c023, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 508.474946] env[62569]: DEBUG nova.compute.utils [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 508.480162] env[62569]: DEBUG nova.compute.manager [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 508.480346] env[62569]: DEBUG nova.network.neutron [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 508.528675] env[62569]: DEBUG nova.policy [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd42f84d78f4a4548af83b23f8612f4e3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a1c80315678a4d37b2a00a372242e0a8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 508.581562] env[62569]: INFO nova.compute.manager [-] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Took 1.05 seconds to deallocate network for instance. [ 508.583085] env[62569]: DEBUG nova.compute.claims [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 508.583408] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 508.730389] env[62569]: DEBUG nova.network.neutron [req-17054613-6a5b-466f-85a4-35769530fe9c req-166526cf-1980-41a0-93f5-99858e878b48 service nova] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 508.852061] env[62569]: DEBUG nova.network.neutron [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Successfully created port: 477c5934-fdf4-446e-82a2-25a80ca82caf {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 508.923223] env[62569]: DEBUG nova.network.neutron [req-17054613-6a5b-466f-85a4-35769530fe9c req-166526cf-1980-41a0-93f5-99858e878b48 service nova] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 508.945299] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 508.945299] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 508.945299] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 508.945299] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 508.945780] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 508.946239] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-11915ea2-83e7-4fdb-9fc7-fe2ea3720ce5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.964589] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 508.964968] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 508.965767] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22499980-4f77-4a2b-96ad-eaae49ba4f54 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.972668] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e213633-019f-4726-960d-9a519c174235 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.978439] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Waiting for the task: (returnval){ [ 508.978439] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52679183-8ce4-a532-02e8-b231225f0dd2" [ 508.978439] env[62569]: _type = "Task" [ 508.978439] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 508.988731] env[62569]: DEBUG nova.compute.manager [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 508.991513] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52679183-8ce4-a532-02e8-b231225f0dd2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 509.149811] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d73c4405-72be-4220-9e54-5c8d0ed20301 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.159105] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b57b6e71-1700-4e9c-862e-b1bdbc68d49e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.199772] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-811778f1-5a72-4da1-80b2-b0c635333257 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.208341] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14e5f536-501e-46eb-88ee-145cb239751f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.224477] env[62569]: DEBUG nova.compute.provider_tree [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 509.426350] env[62569]: DEBUG oslo_concurrency.lockutils [req-17054613-6a5b-466f-85a4-35769530fe9c req-166526cf-1980-41a0-93f5-99858e878b48 service nova] Releasing lock "refresh_cache-99412ace-3825-4c62-ad54-e59b83218387" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 509.426643] env[62569]: DEBUG nova.compute.manager [req-17054613-6a5b-466f-85a4-35769530fe9c req-166526cf-1980-41a0-93f5-99858e878b48 service nova] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Received event network-vif-deleted-8d597f38-b911-472d-8ad8-4a66203f2e55 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 509.490753] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Preparing fetch location {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 509.491435] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Creating directory with path [datastore1] vmware_temp/7ad3fcbb-d4f1-40c8-9461-d52a56a49daf/f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 509.491766] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-77d86d9b-a320-45bd-b5a4-80462db6bb1e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.518310] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Created directory with path [datastore1] vmware_temp/7ad3fcbb-d4f1-40c8-9461-d52a56a49daf/f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 509.518310] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Fetch image to [datastore1] vmware_temp/7ad3fcbb-d4f1-40c8-9461-d52a56a49daf/f77800cf-af0d-4e9c-b312-2c59488f2c7c/tmp-sparse.vmdk {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 509.518487] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Downloading image file data f77800cf-af0d-4e9c-b312-2c59488f2c7c to [datastore1] vmware_temp/7ad3fcbb-d4f1-40c8-9461-d52a56a49daf/f77800cf-af0d-4e9c-b312-2c59488f2c7c/tmp-sparse.vmdk on the data store datastore1 {{(pid=62569) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 509.519503] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-520fc7ed-7f39-4fca-a772-8783f8e3e92f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.528971] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08a10f60-9188-4e1f-a4f5-4c913c99795d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.543500] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8ab7dae-0d9a-4b0c-914b-501a8714e337 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.580503] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53989200-9ebd-48cc-ae0b-085201b7e6ab {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.587285] env[62569]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-669eb09b-1880-4204-9a96-8094f8fd4c86 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.636365] env[62569]: DEBUG oslo_concurrency.lockutils [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Acquiring lock "597c9673-bbd4-4dd0-829a-c05f8667c951" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 509.636695] env[62569]: DEBUG oslo_concurrency.lockutils [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Lock "597c9673-bbd4-4dd0-829a-c05f8667c951" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 509.680231] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Downloading image file data f77800cf-af0d-4e9c-b312-2c59488f2c7c to the data store datastore1 {{(pid=62569) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 509.731932] env[62569]: DEBUG nova.scheduler.client.report [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 509.824784] env[62569]: DEBUG oslo_vmware.rw_handles [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/7ad3fcbb-d4f1-40c8-9461-d52a56a49daf/f77800cf-af0d-4e9c-b312-2c59488f2c7c/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62569) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 509.999342] env[62569]: DEBUG nova.compute.manager [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 510.140274] env[62569]: DEBUG nova.compute.manager [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 510.240499] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.271s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 510.241084] env[62569]: DEBUG nova.compute.manager [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 510.245556] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.610s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.249395] env[62569]: INFO nova.compute.claims [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 510.317763] env[62569]: DEBUG nova.virt.hardware [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 510.318027] env[62569]: DEBUG nova.virt.hardware [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 510.318188] env[62569]: DEBUG nova.virt.hardware [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 510.320540] env[62569]: DEBUG nova.virt.hardware [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 510.320540] env[62569]: DEBUG nova.virt.hardware [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 510.320540] env[62569]: DEBUG nova.virt.hardware [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 510.320540] env[62569]: DEBUG nova.virt.hardware [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 510.320540] env[62569]: DEBUG nova.virt.hardware [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 510.320832] env[62569]: DEBUG nova.virt.hardware [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 510.320832] env[62569]: DEBUG nova.virt.hardware [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 510.320832] env[62569]: DEBUG nova.virt.hardware [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 510.320832] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2924441c-636c-4b82-b7eb-41042c21aaaf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.332957] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-947b946d-e179-4bad-ba99-6fc94a54955c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.628566] env[62569]: ERROR nova.compute.manager [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 477c5934-fdf4-446e-82a2-25a80ca82caf, please check neutron logs for more information. [ 510.628566] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 510.628566] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 510.628566] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 510.628566] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 510.628566] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 510.628566] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 510.628566] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 510.628566] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 510.628566] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 510.628566] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 510.628566] env[62569]: ERROR nova.compute.manager raise self.value [ 510.628566] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 510.628566] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 510.628566] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 510.628566] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 510.629079] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 510.629079] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 510.629079] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 477c5934-fdf4-446e-82a2-25a80ca82caf, please check neutron logs for more information. [ 510.629079] env[62569]: ERROR nova.compute.manager [ 510.629079] env[62569]: Traceback (most recent call last): [ 510.629079] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 510.629079] env[62569]: listener.cb(fileno) [ 510.629079] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 510.629079] env[62569]: result = function(*args, **kwargs) [ 510.629079] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 510.629079] env[62569]: return func(*args, **kwargs) [ 510.629079] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 510.629079] env[62569]: raise e [ 510.629079] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 510.629079] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 510.629079] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 510.629079] env[62569]: created_port_ids = self._update_ports_for_instance( [ 510.629079] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 510.629079] env[62569]: with excutils.save_and_reraise_exception(): [ 510.629079] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 510.629079] env[62569]: self.force_reraise() [ 510.629079] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 510.629079] env[62569]: raise self.value [ 510.629079] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 510.629079] env[62569]: updated_port = self._update_port( [ 510.629079] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 510.629079] env[62569]: _ensure_no_port_binding_failure(port) [ 510.629079] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 510.629079] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 510.629870] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 477c5934-fdf4-446e-82a2-25a80ca82caf, please check neutron logs for more information. [ 510.629870] env[62569]: Removing descriptor: 14 [ 510.629870] env[62569]: ERROR nova.compute.manager [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 477c5934-fdf4-446e-82a2-25a80ca82caf, please check neutron logs for more information. [ 510.629870] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Traceback (most recent call last): [ 510.629870] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 510.629870] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] yield resources [ 510.629870] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 510.629870] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] self.driver.spawn(context, instance, image_meta, [ 510.629870] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 510.629870] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] self._vmops.spawn(context, instance, image_meta, injected_files, [ 510.629870] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 510.629870] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] vm_ref = self.build_virtual_machine(instance, [ 510.630174] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 510.630174] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] vif_infos = vmwarevif.get_vif_info(self._session, [ 510.630174] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 510.630174] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] for vif in network_info: [ 510.630174] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 510.630174] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] return self._sync_wrapper(fn, *args, **kwargs) [ 510.630174] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 510.630174] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] self.wait() [ 510.630174] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 510.630174] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] self[:] = self._gt.wait() [ 510.630174] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 510.630174] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] return self._exit_event.wait() [ 510.630174] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 510.630640] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] result = hub.switch() [ 510.630640] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 510.630640] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] return self.greenlet.switch() [ 510.630640] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 510.630640] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] result = function(*args, **kwargs) [ 510.630640] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 510.630640] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] return func(*args, **kwargs) [ 510.630640] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 510.630640] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] raise e [ 510.630640] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 510.630640] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] nwinfo = self.network_api.allocate_for_instance( [ 510.630640] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 510.630640] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] created_port_ids = self._update_ports_for_instance( [ 510.630983] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 510.630983] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] with excutils.save_and_reraise_exception(): [ 510.630983] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 510.630983] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] self.force_reraise() [ 510.630983] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 510.630983] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] raise self.value [ 510.630983] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 510.630983] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] updated_port = self._update_port( [ 510.630983] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 510.630983] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] _ensure_no_port_binding_failure(port) [ 510.630983] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 510.630983] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] raise exception.PortBindingFailed(port_id=port['id']) [ 510.631392] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] nova.exception.PortBindingFailed: Binding failed for port 477c5934-fdf4-446e-82a2-25a80ca82caf, please check neutron logs for more information. [ 510.631392] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] [ 510.631392] env[62569]: INFO nova.compute.manager [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Terminating instance [ 510.678135] env[62569]: DEBUG oslo_concurrency.lockutils [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 510.754362] env[62569]: DEBUG nova.compute.utils [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 510.761033] env[62569]: DEBUG nova.compute.manager [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 510.761033] env[62569]: DEBUG nova.network.neutron [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 510.810034] env[62569]: DEBUG oslo_vmware.rw_handles [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Completed reading data from the image iterator. {{(pid=62569) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 510.810034] env[62569]: DEBUG oslo_vmware.rw_handles [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Closing write handle for https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/7ad3fcbb-d4f1-40c8-9461-d52a56a49daf/f77800cf-af0d-4e9c-b312-2c59488f2c7c/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62569) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 510.851498] env[62569]: DEBUG nova.policy [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '733d6801a2ad4e99add107ef5215a795', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1b833f826be944d18834d13e2bff4124', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 510.967332] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Downloaded image file data f77800cf-af0d-4e9c-b312-2c59488f2c7c to vmware_temp/7ad3fcbb-d4f1-40c8-9461-d52a56a49daf/f77800cf-af0d-4e9c-b312-2c59488f2c7c/tmp-sparse.vmdk on the data store datastore1 {{(pid=62569) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 510.969113] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Caching image {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 510.969354] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Copying Virtual Disk [datastore1] vmware_temp/7ad3fcbb-d4f1-40c8-9461-d52a56a49daf/f77800cf-af0d-4e9c-b312-2c59488f2c7c/tmp-sparse.vmdk to [datastore1] vmware_temp/7ad3fcbb-d4f1-40c8-9461-d52a56a49daf/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 510.969554] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-33b46484-7854-4895-a0ba-9fe1f4224427 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.978730] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Waiting for the task: (returnval){ [ 510.978730] env[62569]: value = "task-1249790" [ 510.978730] env[62569]: _type = "Task" [ 510.978730] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 510.990475] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249790, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 511.133228] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Acquiring lock "refresh_cache-3a9b16bb-9d58-4f0a-8fab-16f2486a6702" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 511.133462] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Acquired lock "refresh_cache-3a9b16bb-9d58-4f0a-8fab-16f2486a6702" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 511.133588] env[62569]: DEBUG nova.network.neutron [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 511.161373] env[62569]: DEBUG nova.compute.manager [req-efb29e6b-b0fc-4fd3-9d93-3afe990f3ac1 req-99274d32-558e-4ed7-927b-f265d6efe9f2 service nova] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Received event network-changed-477c5934-fdf4-446e-82a2-25a80ca82caf {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 511.161679] env[62569]: DEBUG nova.compute.manager [req-efb29e6b-b0fc-4fd3-9d93-3afe990f3ac1 req-99274d32-558e-4ed7-927b-f265d6efe9f2 service nova] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Refreshing instance network info cache due to event network-changed-477c5934-fdf4-446e-82a2-25a80ca82caf. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 511.161762] env[62569]: DEBUG oslo_concurrency.lockutils [req-efb29e6b-b0fc-4fd3-9d93-3afe990f3ac1 req-99274d32-558e-4ed7-927b-f265d6efe9f2 service nova] Acquiring lock "refresh_cache-3a9b16bb-9d58-4f0a-8fab-16f2486a6702" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 511.235142] env[62569]: DEBUG nova.network.neutron [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Successfully created port: 8e17f59f-041c-407b-bdfa-f054585d93d7 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 511.262818] env[62569]: DEBUG nova.compute.manager [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 511.494257] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249790, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 511.510215] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce0af9c0-3ba8-45ac-ad64-9d604f752b86 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.518892] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236492a9-2382-4575-9d2a-d71662147e77 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.552891] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0652140e-b089-4f9f-b3f9-aaafdc15469c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.561356] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b3182a7-0bb2-46db-93a5-3df641b837a8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.577821] env[62569]: DEBUG nova.compute.provider_tree [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 511.686050] env[62569]: DEBUG nova.network.neutron [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 511.918316] env[62569]: DEBUG nova.network.neutron [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 511.990698] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249790, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.659305} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 511.990952] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Copied Virtual Disk [datastore1] vmware_temp/7ad3fcbb-d4f1-40c8-9461-d52a56a49daf/f77800cf-af0d-4e9c-b312-2c59488f2c7c/tmp-sparse.vmdk to [datastore1] vmware_temp/7ad3fcbb-d4f1-40c8-9461-d52a56a49daf/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 511.991426] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Deleting the datastore file [datastore1] vmware_temp/7ad3fcbb-d4f1-40c8-9461-d52a56a49daf/f77800cf-af0d-4e9c-b312-2c59488f2c7c/tmp-sparse.vmdk {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 511.992185] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-671190e2-dd57-4e15-b54d-7507d3783137 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.999037] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Waiting for the task: (returnval){ [ 511.999037] env[62569]: value = "task-1249791" [ 511.999037] env[62569]: _type = "Task" [ 511.999037] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 512.008595] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249791, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 512.085810] env[62569]: DEBUG nova.scheduler.client.report [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 512.258256] env[62569]: ERROR nova.compute.manager [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8e17f59f-041c-407b-bdfa-f054585d93d7, please check neutron logs for more information. [ 512.258256] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 512.258256] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 512.258256] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 512.258256] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 512.258256] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 512.258256] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 512.258256] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 512.258256] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 512.258256] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 512.258256] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 512.258256] env[62569]: ERROR nova.compute.manager raise self.value [ 512.258256] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 512.258256] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 512.258256] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 512.258256] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 512.260024] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 512.260024] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 512.260024] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8e17f59f-041c-407b-bdfa-f054585d93d7, please check neutron logs for more information. [ 512.260024] env[62569]: ERROR nova.compute.manager [ 512.260024] env[62569]: Traceback (most recent call last): [ 512.260024] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 512.260024] env[62569]: listener.cb(fileno) [ 512.260024] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 512.260024] env[62569]: result = function(*args, **kwargs) [ 512.260024] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 512.260024] env[62569]: return func(*args, **kwargs) [ 512.260024] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 512.260024] env[62569]: raise e [ 512.260024] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 512.260024] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 512.260024] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 512.260024] env[62569]: created_port_ids = self._update_ports_for_instance( [ 512.260024] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 512.260024] env[62569]: with excutils.save_and_reraise_exception(): [ 512.260024] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 512.260024] env[62569]: self.force_reraise() [ 512.260024] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 512.260024] env[62569]: raise self.value [ 512.260024] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 512.260024] env[62569]: updated_port = self._update_port( [ 512.260024] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 512.260024] env[62569]: _ensure_no_port_binding_failure(port) [ 512.260024] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 512.260024] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 512.261343] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 8e17f59f-041c-407b-bdfa-f054585d93d7, please check neutron logs for more information. [ 512.261343] env[62569]: Removing descriptor: 14 [ 512.278318] env[62569]: DEBUG nova.compute.manager [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 512.309883] env[62569]: DEBUG nova.virt.hardware [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 512.310049] env[62569]: DEBUG nova.virt.hardware [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 512.310259] env[62569]: DEBUG nova.virt.hardware [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 512.310476] env[62569]: DEBUG nova.virt.hardware [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 512.310632] env[62569]: DEBUG nova.virt.hardware [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 512.314325] env[62569]: DEBUG nova.virt.hardware [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 512.314575] env[62569]: DEBUG nova.virt.hardware [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 512.314946] env[62569]: DEBUG nova.virt.hardware [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 512.314946] env[62569]: DEBUG nova.virt.hardware [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 512.315077] env[62569]: DEBUG nova.virt.hardware [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 512.315405] env[62569]: DEBUG nova.virt.hardware [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 512.316164] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff625feb-7d9f-41d8-9aae-b8dcf53f6f52 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.325236] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2454afa5-d424-4ca8-8edf-cf2cf3f3e876 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.345615] env[62569]: ERROR nova.compute.manager [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8e17f59f-041c-407b-bdfa-f054585d93d7, please check neutron logs for more information. [ 512.345615] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Traceback (most recent call last): [ 512.345615] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 512.345615] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] yield resources [ 512.345615] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 512.345615] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] self.driver.spawn(context, instance, image_meta, [ 512.345615] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 512.345615] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] self._vmops.spawn(context, instance, image_meta, injected_files, [ 512.345615] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 512.345615] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] vm_ref = self.build_virtual_machine(instance, [ 512.345615] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 512.346027] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] vif_infos = vmwarevif.get_vif_info(self._session, [ 512.346027] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 512.346027] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] for vif in network_info: [ 512.346027] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 512.346027] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] return self._sync_wrapper(fn, *args, **kwargs) [ 512.346027] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 512.346027] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] self.wait() [ 512.346027] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 512.346027] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] self[:] = self._gt.wait() [ 512.346027] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 512.346027] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] return self._exit_event.wait() [ 512.346027] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 512.346027] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] current.throw(*self._exc) [ 512.346402] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 512.346402] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] result = function(*args, **kwargs) [ 512.346402] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 512.346402] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] return func(*args, **kwargs) [ 512.346402] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 512.346402] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] raise e [ 512.346402] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 512.346402] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] nwinfo = self.network_api.allocate_for_instance( [ 512.346402] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 512.346402] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] created_port_ids = self._update_ports_for_instance( [ 512.346402] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 512.346402] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] with excutils.save_and_reraise_exception(): [ 512.346402] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 512.346807] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] self.force_reraise() [ 512.346807] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 512.346807] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] raise self.value [ 512.346807] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 512.346807] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] updated_port = self._update_port( [ 512.346807] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 512.346807] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] _ensure_no_port_binding_failure(port) [ 512.346807] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 512.346807] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] raise exception.PortBindingFailed(port_id=port['id']) [ 512.346807] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] nova.exception.PortBindingFailed: Binding failed for port 8e17f59f-041c-407b-bdfa-f054585d93d7, please check neutron logs for more information. [ 512.346807] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] [ 512.346807] env[62569]: INFO nova.compute.manager [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Terminating instance [ 512.421415] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Releasing lock "refresh_cache-3a9b16bb-9d58-4f0a-8fab-16f2486a6702" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 512.422350] env[62569]: DEBUG nova.compute.manager [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 512.422560] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 512.422880] env[62569]: DEBUG oslo_concurrency.lockutils [req-efb29e6b-b0fc-4fd3-9d93-3afe990f3ac1 req-99274d32-558e-4ed7-927b-f265d6efe9f2 service nova] Acquired lock "refresh_cache-3a9b16bb-9d58-4f0a-8fab-16f2486a6702" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 512.423066] env[62569]: DEBUG nova.network.neutron [req-efb29e6b-b0fc-4fd3-9d93-3afe990f3ac1 req-99274d32-558e-4ed7-927b-f265d6efe9f2 service nova] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Refreshing network info cache for port 477c5934-fdf4-446e-82a2-25a80ca82caf {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 512.425493] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e1f98d38-18c9-47ac-a360-28e7e1f185b2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.437744] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffecc487-39c1-48db-80ab-57717456be7e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.461292] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3a9b16bb-9d58-4f0a-8fab-16f2486a6702 could not be found. [ 512.461420] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 512.461679] env[62569]: INFO nova.compute.manager [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Took 0.04 seconds to destroy the instance on the hypervisor. [ 512.461918] env[62569]: DEBUG oslo.service.loopingcall [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 512.462063] env[62569]: DEBUG nova.compute.manager [-] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 512.462156] env[62569]: DEBUG nova.network.neutron [-] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 512.489930] env[62569]: DEBUG nova.network.neutron [-] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 512.515877] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249791, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.026766} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 512.515877] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 512.515877] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Moving file from [datastore1] vmware_temp/7ad3fcbb-d4f1-40c8-9461-d52a56a49daf/f77800cf-af0d-4e9c-b312-2c59488f2c7c to [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c. {{(pid=62569) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 512.516209] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-8d77f14b-6cda-4b32-98fe-add5b1d5ed96 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.523826] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Waiting for the task: (returnval){ [ 512.523826] env[62569]: value = "task-1249792" [ 512.523826] env[62569]: _type = "Task" [ 512.523826] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 512.533346] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249792, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 512.593970] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.348s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 512.595029] env[62569]: DEBUG nova.compute.manager [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 512.601600] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 7.839s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 512.853664] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Acquiring lock "refresh_cache-611d4cfa-a888-47bc-8857-f061046a77ed" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 512.853918] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Acquired lock "refresh_cache-611d4cfa-a888-47bc-8857-f061046a77ed" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 512.854049] env[62569]: DEBUG nova.network.neutron [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 512.985727] env[62569]: DEBUG nova.network.neutron [req-efb29e6b-b0fc-4fd3-9d93-3afe990f3ac1 req-99274d32-558e-4ed7-927b-f265d6efe9f2 service nova] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 512.996808] env[62569]: DEBUG nova.network.neutron [-] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 513.038944] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249792, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.024404} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 513.039274] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] File moved {{(pid=62569) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 513.039471] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Cleaning up location [datastore1] vmware_temp/7ad3fcbb-d4f1-40c8-9461-d52a56a49daf {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 513.039633] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Deleting the datastore file [datastore1] vmware_temp/7ad3fcbb-d4f1-40c8-9461-d52a56a49daf {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 513.039891] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d1442fc8-5e3e-4989-9c73-21d79ca5d2cb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.051385] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Waiting for the task: (returnval){ [ 513.051385] env[62569]: value = "task-1249793" [ 513.051385] env[62569]: _type = "Task" [ 513.051385] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 513.063923] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249793, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 513.105493] env[62569]: DEBUG nova.compute.utils [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 513.111171] env[62569]: DEBUG nova.compute.manager [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 513.111350] env[62569]: DEBUG nova.network.neutron [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 513.269648] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5980c87-271e-40ef-a18b-2eac34a1d468 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.279605] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9aa1030-02aa-4385-be52-7fc7b629d484 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.317658] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbaea705-6518-495c-b6cb-4c22dc31a3c4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.326796] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f01515-597d-47ef-8a62-3a50ffd1c42a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.345868] env[62569]: DEBUG nova.compute.provider_tree [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 513.372781] env[62569]: DEBUG nova.network.neutron [req-efb29e6b-b0fc-4fd3-9d93-3afe990f3ac1 req-99274d32-558e-4ed7-927b-f265d6efe9f2 service nova] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 513.420466] env[62569]: DEBUG nova.network.neutron [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 513.500485] env[62569]: INFO nova.compute.manager [-] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Took 1.04 seconds to deallocate network for instance. [ 513.504431] env[62569]: DEBUG nova.compute.claims [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 513.504622] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.521513] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Acquiring lock "c542f5dd-10d8-4959-a3ab-eab9b8f4525a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.521513] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Lock "c542f5dd-10d8-4959-a3ab-eab9b8f4525a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 513.542097] env[62569]: DEBUG nova.policy [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cd221453e5a249fda0584b814cd4d92a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd42c337b072e469db9ad895d6b396b7b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 513.562503] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249793, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.096524} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 513.562945] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 513.563666] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-014fd2ca-30af-41b7-9853-47b474a373ca {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.569389] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Waiting for the task: (returnval){ [ 513.569389] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ab9f26-caea-1bd6-713c-405d7f9fef4f" [ 513.569389] env[62569]: _type = "Task" [ 513.569389] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 513.581095] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ab9f26-caea-1bd6-713c-405d7f9fef4f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 513.604821] env[62569]: DEBUG nova.network.neutron [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 513.612295] env[62569]: DEBUG nova.compute.manager [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 513.850646] env[62569]: DEBUG nova.scheduler.client.report [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 513.876935] env[62569]: DEBUG oslo_concurrency.lockutils [req-efb29e6b-b0fc-4fd3-9d93-3afe990f3ac1 req-99274d32-558e-4ed7-927b-f265d6efe9f2 service nova] Releasing lock "refresh_cache-3a9b16bb-9d58-4f0a-8fab-16f2486a6702" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 514.023702] env[62569]: DEBUG nova.compute.manager [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 514.093068] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ab9f26-caea-1bd6-713c-405d7f9fef4f, 'name': SearchDatastore_Task, 'duration_secs': 0.038889} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 514.093394] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 514.094843] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007/17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 514.094843] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7a144006-0630-4a26-810e-835f5ee9788e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.103879] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Waiting for the task: (returnval){ [ 514.103879] env[62569]: value = "task-1249794" [ 514.103879] env[62569]: _type = "Task" [ 514.103879] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 514.106869] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Releasing lock "refresh_cache-611d4cfa-a888-47bc-8857-f061046a77ed" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 514.107400] env[62569]: DEBUG nova.compute.manager [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 514.107737] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 514.107908] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1c253241-0e34-4ede-837e-66479d231abb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.124756] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249794, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 514.140110] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e992775-4fa1-4375-b3fc-618b13e9edaa {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.175894] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 611d4cfa-a888-47bc-8857-f061046a77ed could not be found. [ 514.175894] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 514.175894] env[62569]: INFO nova.compute.manager [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Took 0.07 seconds to destroy the instance on the hypervisor. [ 514.175894] env[62569]: DEBUG oslo.service.loopingcall [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 514.176150] env[62569]: DEBUG nova.compute.manager [-] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 514.176150] env[62569]: DEBUG nova.network.neutron [-] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 514.216684] env[62569]: DEBUG nova.network.neutron [-] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 514.357629] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.757s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 514.357983] env[62569]: ERROR nova.compute.manager [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6abfdbf1-8e5c-499b-98d7-789b91c9ee3a, please check neutron logs for more information. [ 514.357983] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Traceback (most recent call last): [ 514.357983] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 514.357983] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] self.driver.spawn(context, instance, image_meta, [ 514.357983] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 514.357983] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] self._vmops.spawn(context, instance, image_meta, injected_files, [ 514.357983] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 514.357983] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] vm_ref = self.build_virtual_machine(instance, [ 514.357983] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 514.357983] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] vif_infos = vmwarevif.get_vif_info(self._session, [ 514.357983] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 514.358401] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] for vif in network_info: [ 514.358401] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 514.358401] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] return self._sync_wrapper(fn, *args, **kwargs) [ 514.358401] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 514.358401] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] self.wait() [ 514.358401] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 514.358401] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] self[:] = self._gt.wait() [ 514.358401] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 514.358401] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] return self._exit_event.wait() [ 514.358401] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 514.358401] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] result = hub.switch() [ 514.358401] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 514.358401] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] return self.greenlet.switch() [ 514.358798] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 514.358798] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] result = function(*args, **kwargs) [ 514.358798] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 514.358798] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] return func(*args, **kwargs) [ 514.358798] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 514.358798] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] raise e [ 514.358798] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 514.358798] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] nwinfo = self.network_api.allocate_for_instance( [ 514.358798] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 514.358798] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] created_port_ids = self._update_ports_for_instance( [ 514.358798] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 514.358798] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] with excutils.save_and_reraise_exception(): [ 514.358798] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 514.360048] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] self.force_reraise() [ 514.360048] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 514.360048] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] raise self.value [ 514.360048] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 514.360048] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] updated_port = self._update_port( [ 514.360048] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 514.360048] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] _ensure_no_port_binding_failure(port) [ 514.360048] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 514.360048] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] raise exception.PortBindingFailed(port_id=port['id']) [ 514.360048] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] nova.exception.PortBindingFailed: Binding failed for port 6abfdbf1-8e5c-499b-98d7-789b91c9ee3a, please check neutron logs for more information. [ 514.360048] env[62569]: ERROR nova.compute.manager [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] [ 514.360347] env[62569]: DEBUG nova.compute.utils [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Binding failed for port 6abfdbf1-8e5c-499b-98d7-789b91c9ee3a, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 514.360347] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.273s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 514.363281] env[62569]: INFO nova.compute.claims [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 514.374431] env[62569]: DEBUG nova.compute.manager [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Build of instance a2d5b36d-357f-42b0-8538-77c5ef915d61 was re-scheduled: Binding failed for port 6abfdbf1-8e5c-499b-98d7-789b91c9ee3a, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 514.374431] env[62569]: DEBUG nova.compute.manager [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 514.374431] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Acquiring lock "refresh_cache-a2d5b36d-357f-42b0-8538-77c5ef915d61" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 514.374431] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Acquired lock "refresh_cache-a2d5b36d-357f-42b0-8538-77c5ef915d61" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 514.374683] env[62569]: DEBUG nova.network.neutron [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 514.569510] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.623463] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249794, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 514.629287] env[62569]: DEBUG nova.compute.manager [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 514.660205] env[62569]: DEBUG nova.virt.hardware [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 514.663626] env[62569]: DEBUG nova.virt.hardware [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 514.663626] env[62569]: DEBUG nova.virt.hardware [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 514.663626] env[62569]: DEBUG nova.virt.hardware [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 514.663626] env[62569]: DEBUG nova.virt.hardware [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 514.663626] env[62569]: DEBUG nova.virt.hardware [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 514.663792] env[62569]: DEBUG nova.virt.hardware [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 514.664837] env[62569]: DEBUG nova.virt.hardware [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 514.665533] env[62569]: DEBUG nova.virt.hardware [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 514.665533] env[62569]: DEBUG nova.virt.hardware [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 514.665642] env[62569]: DEBUG nova.virt.hardware [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 514.666899] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82397512-3465-4cfa-9857-cd924b4825ae {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.674786] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7209f99-e1d0-40fe-9f8f-2844e684ef16 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.722689] env[62569]: DEBUG nova.network.neutron [-] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 514.796913] env[62569]: DEBUG nova.compute.manager [req-de9998b5-5695-4040-9af3-329945c77050 req-756b1a24-8521-49f6-bc96-c287e81047d9 service nova] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Received event network-vif-deleted-477c5934-fdf4-446e-82a2-25a80ca82caf {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 514.797194] env[62569]: DEBUG nova.compute.manager [req-de9998b5-5695-4040-9af3-329945c77050 req-756b1a24-8521-49f6-bc96-c287e81047d9 service nova] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Received event network-changed-8e17f59f-041c-407b-bdfa-f054585d93d7 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 514.797356] env[62569]: DEBUG nova.compute.manager [req-de9998b5-5695-4040-9af3-329945c77050 req-756b1a24-8521-49f6-bc96-c287e81047d9 service nova] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Refreshing instance network info cache due to event network-changed-8e17f59f-041c-407b-bdfa-f054585d93d7. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 514.797564] env[62569]: DEBUG oslo_concurrency.lockutils [req-de9998b5-5695-4040-9af3-329945c77050 req-756b1a24-8521-49f6-bc96-c287e81047d9 service nova] Acquiring lock "refresh_cache-611d4cfa-a888-47bc-8857-f061046a77ed" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 514.797692] env[62569]: DEBUG oslo_concurrency.lockutils [req-de9998b5-5695-4040-9af3-329945c77050 req-756b1a24-8521-49f6-bc96-c287e81047d9 service nova] Acquired lock "refresh_cache-611d4cfa-a888-47bc-8857-f061046a77ed" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 514.797841] env[62569]: DEBUG nova.network.neutron [req-de9998b5-5695-4040-9af3-329945c77050 req-756b1a24-8521-49f6-bc96-c287e81047d9 service nova] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Refreshing network info cache for port 8e17f59f-041c-407b-bdfa-f054585d93d7 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 514.937985] env[62569]: DEBUG nova.network.neutron [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 515.118103] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249794, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.58294} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 515.118103] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007/17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 515.118103] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 515.118375] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b312c7a1-5d3b-4d59-8b2e-d38d9972f003 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.126137] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Waiting for the task: (returnval){ [ 515.126137] env[62569]: value = "task-1249795" [ 515.126137] env[62569]: _type = "Task" [ 515.126137] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 515.136809] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249795, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 515.138061] env[62569]: DEBUG nova.network.neutron [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 515.228173] env[62569]: INFO nova.compute.manager [-] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Took 1.05 seconds to deallocate network for instance. [ 515.231987] env[62569]: DEBUG nova.compute.claims [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 515.231987] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.341192] env[62569]: DEBUG nova.network.neutron [req-de9998b5-5695-4040-9af3-329945c77050 req-756b1a24-8521-49f6-bc96-c287e81047d9 service nova] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 515.463139] env[62569]: DEBUG nova.network.neutron [req-de9998b5-5695-4040-9af3-329945c77050 req-756b1a24-8521-49f6-bc96-c287e81047d9 service nova] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 515.583158] env[62569]: DEBUG nova.network.neutron [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Successfully created port: 409365e0-c0a3-4367-add1-3803636b9262 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 515.596697] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afb954e4-752b-49dd-a187-c1a78000a041 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.604217] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-513e8781-3c18-4eb4-9b69-7346989523a9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.643920] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a35122-e0bb-4526-ada8-2e0c160d913b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.646908] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Releasing lock "refresh_cache-a2d5b36d-357f-42b0-8538-77c5ef915d61" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 515.647204] env[62569]: DEBUG nova.compute.manager [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 515.647406] env[62569]: DEBUG nova.compute.manager [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 515.647622] env[62569]: DEBUG nova.network.neutron [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 515.654263] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249795, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.177367} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 515.657053] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 515.658017] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e5bbb74-9da3-4208-87cd-6551d3672856 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.662588] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd8e02d-ca5a-4d4a-8849-ceb88c698271 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.679818] env[62569]: DEBUG nova.compute.provider_tree [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 515.701417] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Reconfiguring VM instance instance-00000003 to attach disk [datastore1] 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007/17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 515.702865] env[62569]: DEBUG nova.network.neutron [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 515.704449] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8c5b230-f5fe-43e8-9e5f-56b5047660bc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.720916] env[62569]: DEBUG nova.network.neutron [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 515.728037] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Waiting for the task: (returnval){ [ 515.728037] env[62569]: value = "task-1249796" [ 515.728037] env[62569]: _type = "Task" [ 515.728037] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 515.737654] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249796, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 515.966248] env[62569]: DEBUG oslo_concurrency.lockutils [req-de9998b5-5695-4040-9af3-329945c77050 req-756b1a24-8521-49f6-bc96-c287e81047d9 service nova] Releasing lock "refresh_cache-611d4cfa-a888-47bc-8857-f061046a77ed" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 515.966513] env[62569]: DEBUG nova.compute.manager [req-de9998b5-5695-4040-9af3-329945c77050 req-756b1a24-8521-49f6-bc96-c287e81047d9 service nova] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Received event network-vif-deleted-8e17f59f-041c-407b-bdfa-f054585d93d7 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 516.204917] env[62569]: DEBUG nova.scheduler.client.report [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 516.222964] env[62569]: INFO nova.compute.manager [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: a2d5b36d-357f-42b0-8538-77c5ef915d61] Took 0.58 seconds to deallocate network for instance. [ 516.241169] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249796, 'name': ReconfigVM_Task, 'duration_secs': 0.315318} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 516.241330] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Reconfigured VM instance instance-00000003 to attach disk [datastore1] 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007/17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 516.242026] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5d3a4499-25d7-4c28-ab75-80e6fff66483 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.250517] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Waiting for the task: (returnval){ [ 516.250517] env[62569]: value = "task-1249797" [ 516.250517] env[62569]: _type = "Task" [ 516.250517] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 516.260472] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249797, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 516.709927] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.350s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 516.710531] env[62569]: DEBUG nova.compute.manager [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 516.716867] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.133s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.767826] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249797, 'name': Rename_Task, 'duration_secs': 0.134351} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 516.767826] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 516.767826] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a00e9468-78a1-41fd-b165-286819ea9d9d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.775140] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Waiting for the task: (returnval){ [ 516.775140] env[62569]: value = "task-1249798" [ 516.775140] env[62569]: _type = "Task" [ 516.775140] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 516.788910] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249798, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 517.221443] env[62569]: DEBUG nova.compute.utils [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 517.230591] env[62569]: DEBUG nova.compute.manager [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 517.230591] env[62569]: DEBUG nova.network.neutron [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 517.289214] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249798, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 517.298773] env[62569]: INFO nova.scheduler.client.report [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Deleted allocations for instance a2d5b36d-357f-42b0-8538-77c5ef915d61 [ 517.465579] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5933ad1a-2e83-4f5e-b75a-12a785eb0730 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.477957] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d26543-83e2-4013-b0e4-2a0ab02c4c45 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.518264] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-009d186c-f727-4f24-af70-1718debef722 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.527611] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c32cb42-c4f7-47b3-956f-e4611a19def2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.542300] env[62569]: DEBUG nova.compute.provider_tree [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 517.586264] env[62569]: DEBUG nova.policy [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '840341f9ca2f4eb5b73a926798c5be8f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd38467713b6340ab9ca2e0d8cbd799a6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 517.730783] env[62569]: DEBUG nova.compute.manager [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 517.781427] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Acquiring lock "ba23eea7-422c-4e5d-b95e-f95589e41550" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.781694] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Lock "ba23eea7-422c-4e5d-b95e-f95589e41550" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.804428] env[62569]: DEBUG oslo_vmware.api [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249798, 'name': PowerOnVM_Task, 'duration_secs': 0.673325} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 517.804705] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 517.804908] env[62569]: INFO nova.compute.manager [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Took 10.04 seconds to spawn the instance on the hypervisor. [ 517.805189] env[62569]: DEBUG nova.compute.manager [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 517.806113] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9aec467-e202-471b-9f13-01ec7d74eb70 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.814212] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cb1872e-acef-4613-a1ed-7324e9528f75 tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Lock "a2d5b36d-357f-42b0-8538-77c5ef915d61" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.149s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 518.045567] env[62569]: DEBUG nova.scheduler.client.report [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 518.290975] env[62569]: DEBUG nova.compute.manager [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 518.336750] env[62569]: INFO nova.compute.manager [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Took 16.86 seconds to build instance. [ 518.554780] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.838s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 518.555310] env[62569]: ERROR nova.compute.manager [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8d597f38-b911-472d-8ad8-4a66203f2e55, please check neutron logs for more information. [ 518.555310] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] Traceback (most recent call last): [ 518.555310] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 518.555310] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] self.driver.spawn(context, instance, image_meta, [ 518.555310] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 518.555310] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] self._vmops.spawn(context, instance, image_meta, injected_files, [ 518.555310] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 518.555310] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] vm_ref = self.build_virtual_machine(instance, [ 518.555310] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 518.555310] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] vif_infos = vmwarevif.get_vif_info(self._session, [ 518.555310] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 518.555642] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] for vif in network_info: [ 518.555642] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 518.555642] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] return self._sync_wrapper(fn, *args, **kwargs) [ 518.555642] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 518.555642] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] self.wait() [ 518.555642] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 518.555642] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] self[:] = self._gt.wait() [ 518.555642] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 518.555642] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] return self._exit_event.wait() [ 518.555642] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 518.555642] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] result = hub.switch() [ 518.555642] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 518.555642] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] return self.greenlet.switch() [ 518.555958] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 518.555958] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] result = function(*args, **kwargs) [ 518.555958] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 518.555958] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] return func(*args, **kwargs) [ 518.555958] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 518.555958] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] raise e [ 518.555958] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 518.555958] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] nwinfo = self.network_api.allocate_for_instance( [ 518.555958] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 518.555958] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] created_port_ids = self._update_ports_for_instance( [ 518.555958] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 518.555958] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] with excutils.save_and_reraise_exception(): [ 518.555958] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 518.556288] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] self.force_reraise() [ 518.556288] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 518.556288] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] raise self.value [ 518.556288] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 518.556288] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] updated_port = self._update_port( [ 518.556288] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 518.556288] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] _ensure_no_port_binding_failure(port) [ 518.556288] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 518.556288] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] raise exception.PortBindingFailed(port_id=port['id']) [ 518.556288] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] nova.exception.PortBindingFailed: Binding failed for port 8d597f38-b911-472d-8ad8-4a66203f2e55, please check neutron logs for more information. [ 518.556288] env[62569]: ERROR nova.compute.manager [instance: 99412ace-3825-4c62-ad54-e59b83218387] [ 518.556585] env[62569]: DEBUG nova.compute.utils [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Binding failed for port 8d597f38-b911-472d-8ad8-4a66203f2e55, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 518.558330] env[62569]: DEBUG oslo_concurrency.lockutils [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.879s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.558943] env[62569]: INFO nova.compute.claims [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 518.562027] env[62569]: DEBUG nova.compute.manager [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Build of instance 99412ace-3825-4c62-ad54-e59b83218387 was re-scheduled: Binding failed for port 8d597f38-b911-472d-8ad8-4a66203f2e55, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 518.562792] env[62569]: DEBUG nova.compute.manager [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 518.562925] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Acquiring lock "refresh_cache-99412ace-3825-4c62-ad54-e59b83218387" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 518.563785] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Acquired lock "refresh_cache-99412ace-3825-4c62-ad54-e59b83218387" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 518.563785] env[62569]: DEBUG nova.network.neutron [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 518.743699] env[62569]: DEBUG nova.compute.manager [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 518.787041] env[62569]: DEBUG nova.virt.hardware [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 518.787240] env[62569]: DEBUG nova.virt.hardware [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 518.787390] env[62569]: DEBUG nova.virt.hardware [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 518.787563] env[62569]: DEBUG nova.virt.hardware [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 518.787699] env[62569]: DEBUG nova.virt.hardware [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 518.787863] env[62569]: DEBUG nova.virt.hardware [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 518.788052] env[62569]: DEBUG nova.virt.hardware [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 518.788208] env[62569]: DEBUG nova.virt.hardware [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 518.788369] env[62569]: DEBUG nova.virt.hardware [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 518.788524] env[62569]: DEBUG nova.virt.hardware [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 518.788682] env[62569]: DEBUG nova.virt.hardware [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 518.789609] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46472ef9-15e8-4808-ac5d-d75e5db69a2e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.800027] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-239552bd-69cf-4c57-bc82-f0a9baa11e79 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.829015] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.842547] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4be89a1d-7909-440b-abb3-64796ba31fc1 tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Lock "17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.378s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 519.114331] env[62569]: DEBUG nova.network.neutron [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 519.343077] env[62569]: DEBUG nova.network.neutron [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Successfully created port: 95c9ce89-4e44-45c7-8dcb-4c285a137ff2 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 519.480204] env[62569]: DEBUG nova.network.neutron [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 519.728017] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b33e347-c7c1-42fc-a1fc-932d20ecb76a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.738702] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b300d023-a90d-406f-b536-24a9937b23ee {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.779373] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03d30a63-cfc0-43f1-b4a2-e1c2b79785fe {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.787801] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04663519-bd32-45a7-8ef6-c7bef6cd445a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.804128] env[62569]: DEBUG nova.compute.provider_tree [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 519.986788] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Releasing lock "refresh_cache-99412ace-3825-4c62-ad54-e59b83218387" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 519.986788] env[62569]: DEBUG nova.compute.manager [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 519.986788] env[62569]: DEBUG nova.compute.manager [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 519.986788] env[62569]: DEBUG nova.network.neutron [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 520.030796] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Acquiring lock "e8b8b029-b20a-4af3-a5dc-1b032ce46f8d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.032402] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Lock "e8b8b029-b20a-4af3-a5dc-1b032ce46f8d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.141277] env[62569]: DEBUG nova.network.neutron [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 520.307496] env[62569]: DEBUG nova.scheduler.client.report [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 520.521730] env[62569]: ERROR nova.compute.manager [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 409365e0-c0a3-4367-add1-3803636b9262, please check neutron logs for more information. [ 520.521730] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 520.521730] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 520.521730] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 520.521730] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 520.521730] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 520.521730] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 520.521730] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 520.521730] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 520.521730] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 520.521730] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 520.521730] env[62569]: ERROR nova.compute.manager raise self.value [ 520.521730] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 520.521730] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 520.521730] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 520.521730] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 520.522456] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 520.522456] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 520.522456] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 409365e0-c0a3-4367-add1-3803636b9262, please check neutron logs for more information. [ 520.522456] env[62569]: ERROR nova.compute.manager [ 520.522456] env[62569]: Traceback (most recent call last): [ 520.522456] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 520.522456] env[62569]: listener.cb(fileno) [ 520.522456] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 520.522456] env[62569]: result = function(*args, **kwargs) [ 520.522456] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 520.522456] env[62569]: return func(*args, **kwargs) [ 520.522456] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 520.522456] env[62569]: raise e [ 520.522456] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 520.522456] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 520.522456] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 520.522456] env[62569]: created_port_ids = self._update_ports_for_instance( [ 520.522456] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 520.522456] env[62569]: with excutils.save_and_reraise_exception(): [ 520.522456] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 520.522456] env[62569]: self.force_reraise() [ 520.522456] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 520.522456] env[62569]: raise self.value [ 520.522456] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 520.522456] env[62569]: updated_port = self._update_port( [ 520.522456] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 520.522456] env[62569]: _ensure_no_port_binding_failure(port) [ 520.522456] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 520.522456] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 520.523706] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 409365e0-c0a3-4367-add1-3803636b9262, please check neutron logs for more information. [ 520.523706] env[62569]: Removing descriptor: 16 [ 520.523706] env[62569]: ERROR nova.compute.manager [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 409365e0-c0a3-4367-add1-3803636b9262, please check neutron logs for more information. [ 520.523706] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Traceback (most recent call last): [ 520.523706] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 520.523706] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] yield resources [ 520.523706] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 520.523706] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] self.driver.spawn(context, instance, image_meta, [ 520.523706] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 520.523706] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 520.523706] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 520.523706] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] vm_ref = self.build_virtual_machine(instance, [ 520.524024] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 520.524024] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] vif_infos = vmwarevif.get_vif_info(self._session, [ 520.524024] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 520.524024] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] for vif in network_info: [ 520.524024] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 520.524024] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] return self._sync_wrapper(fn, *args, **kwargs) [ 520.524024] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 520.524024] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] self.wait() [ 520.524024] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 520.524024] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] self[:] = self._gt.wait() [ 520.524024] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 520.524024] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] return self._exit_event.wait() [ 520.524024] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 520.524394] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] result = hub.switch() [ 520.524394] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 520.524394] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] return self.greenlet.switch() [ 520.524394] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 520.524394] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] result = function(*args, **kwargs) [ 520.524394] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 520.524394] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] return func(*args, **kwargs) [ 520.524394] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 520.524394] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] raise e [ 520.524394] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 520.524394] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] nwinfo = self.network_api.allocate_for_instance( [ 520.524394] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 520.524394] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] created_port_ids = self._update_ports_for_instance( [ 520.524717] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 520.524717] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] with excutils.save_and_reraise_exception(): [ 520.524717] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 520.524717] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] self.force_reraise() [ 520.524717] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 520.524717] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] raise self.value [ 520.524717] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 520.524717] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] updated_port = self._update_port( [ 520.524717] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 520.524717] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] _ensure_no_port_binding_failure(port) [ 520.524717] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 520.524717] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] raise exception.PortBindingFailed(port_id=port['id']) [ 520.525028] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] nova.exception.PortBindingFailed: Binding failed for port 409365e0-c0a3-4367-add1-3803636b9262, please check neutron logs for more information. [ 520.525028] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] [ 520.525028] env[62569]: INFO nova.compute.manager [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Terminating instance [ 520.535083] env[62569]: DEBUG nova.compute.manager [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 520.644411] env[62569]: DEBUG nova.network.neutron [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 520.817465] env[62569]: DEBUG oslo_concurrency.lockutils [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.260s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 520.817895] env[62569]: DEBUG nova.compute.manager [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 520.824927] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 7.318s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.028046] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Acquiring lock "refresh_cache-ff90249f-ae4b-4aeb-91ed-d8a41a942a5d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 521.028046] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Acquired lock "refresh_cache-ff90249f-ae4b-4aeb-91ed-d8a41a942a5d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 521.028046] env[62569]: DEBUG nova.network.neutron [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 521.067952] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.146810] env[62569]: INFO nova.compute.manager [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] [instance: 99412ace-3825-4c62-ad54-e59b83218387] Took 1.16 seconds to deallocate network for instance. [ 521.202029] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 521.202029] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 521.202586] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Starting heal instance info cache {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 521.204018] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Rebuilding the list of instances to heal {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10312}} [ 521.326485] env[62569]: DEBUG nova.compute.utils [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 521.333457] env[62569]: DEBUG nova.compute.manager [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 521.333771] env[62569]: DEBUG nova.network.neutron [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 521.498355] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f073a287-4aa1-40ab-b627-17520695b441 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.508644] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-157d1d24-2089-4dd2-b269-820bec0c5626 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.548328] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-082c8d61-b788-4e85-a5a1-0bc926e5856a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.557056] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79d57b9c-4e3b-4874-9d44-388a92e36ac3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.571296] env[62569]: DEBUG nova.compute.provider_tree [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 521.587949] env[62569]: DEBUG nova.network.neutron [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 521.607347] env[62569]: DEBUG nova.policy [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '61504d5fd5f34ef986d3a08988aed62d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '67f23c4781134b30acacfeabbfb7761f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 521.708109] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Skipping network cache update for instance because it is Building. {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 521.708304] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Skipping network cache update for instance because it is Building. {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 521.708621] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Skipping network cache update for instance because it is Building. {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 521.708814] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Skipping network cache update for instance because it is Building. {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 521.758886] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Acquiring lock "6e35fe1c-05bc-4dcc-b7ec-e9998a647739" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.759825] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Lock "6e35fe1c-05bc-4dcc-b7ec-e9998a647739" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.784894] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "refresh_cache-17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 521.785066] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquired lock "refresh_cache-17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 521.785221] env[62569]: DEBUG nova.network.neutron [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Forcefully refreshing network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 521.785422] env[62569]: DEBUG nova.objects.instance [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lazy-loading 'info_cache' on Instance uuid 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 521.834572] env[62569]: DEBUG nova.compute.manager [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 521.911299] env[62569]: DEBUG nova.network.neutron [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 522.074130] env[62569]: DEBUG nova.scheduler.client.report [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 522.195239] env[62569]: INFO nova.scheduler.client.report [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Deleted allocations for instance 99412ace-3825-4c62-ad54-e59b83218387 [ 522.262972] env[62569]: DEBUG nova.compute.manager [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 522.414434] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Releasing lock "refresh_cache-ff90249f-ae4b-4aeb-91ed-d8a41a942a5d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 522.414966] env[62569]: DEBUG nova.compute.manager [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 522.415084] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 522.415801] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-06058c18-e1a7-4cbc-bf17-6844b9202fbb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.426206] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-401a20a4-e895-4819-bb6f-a14ff8dbca99 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.458481] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ff90249f-ae4b-4aeb-91ed-d8a41a942a5d could not be found. [ 522.458481] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 522.458636] env[62569]: INFO nova.compute.manager [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 522.458855] env[62569]: DEBUG oslo.service.loopingcall [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 522.459132] env[62569]: DEBUG nova.compute.manager [-] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 522.459229] env[62569]: DEBUG nova.network.neutron [-] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 522.515319] env[62569]: DEBUG nova.network.neutron [-] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 522.581919] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.760s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.582955] env[62569]: ERROR nova.compute.manager [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 477c5934-fdf4-446e-82a2-25a80ca82caf, please check neutron logs for more information. [ 522.582955] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Traceback (most recent call last): [ 522.582955] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 522.582955] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] self.driver.spawn(context, instance, image_meta, [ 522.582955] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 522.582955] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] self._vmops.spawn(context, instance, image_meta, injected_files, [ 522.582955] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 522.582955] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] vm_ref = self.build_virtual_machine(instance, [ 522.582955] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 522.582955] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] vif_infos = vmwarevif.get_vif_info(self._session, [ 522.582955] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 522.583345] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] for vif in network_info: [ 522.583345] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 522.583345] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] return self._sync_wrapper(fn, *args, **kwargs) [ 522.583345] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 522.583345] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] self.wait() [ 522.583345] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 522.583345] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] self[:] = self._gt.wait() [ 522.583345] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 522.583345] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] return self._exit_event.wait() [ 522.583345] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 522.583345] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] result = hub.switch() [ 522.583345] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 522.583345] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] return self.greenlet.switch() [ 522.583684] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 522.583684] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] result = function(*args, **kwargs) [ 522.583684] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 522.583684] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] return func(*args, **kwargs) [ 522.583684] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 522.583684] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] raise e [ 522.583684] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 522.583684] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] nwinfo = self.network_api.allocate_for_instance( [ 522.583684] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 522.583684] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] created_port_ids = self._update_ports_for_instance( [ 522.583684] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 522.583684] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] with excutils.save_and_reraise_exception(): [ 522.583684] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 522.584104] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] self.force_reraise() [ 522.584104] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 522.584104] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] raise self.value [ 522.584104] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 522.584104] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] updated_port = self._update_port( [ 522.584104] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 522.584104] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] _ensure_no_port_binding_failure(port) [ 522.584104] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 522.584104] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] raise exception.PortBindingFailed(port_id=port['id']) [ 522.584104] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] nova.exception.PortBindingFailed: Binding failed for port 477c5934-fdf4-446e-82a2-25a80ca82caf, please check neutron logs for more information. [ 522.584104] env[62569]: ERROR nova.compute.manager [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] [ 522.584398] env[62569]: DEBUG nova.compute.utils [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Binding failed for port 477c5934-fdf4-446e-82a2-25a80ca82caf, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 522.584979] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.016s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.588344] env[62569]: INFO nova.compute.claims [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 522.596106] env[62569]: DEBUG nova.compute.manager [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Build of instance 3a9b16bb-9d58-4f0a-8fab-16f2486a6702 was re-scheduled: Binding failed for port 477c5934-fdf4-446e-82a2-25a80ca82caf, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 522.596685] env[62569]: DEBUG nova.compute.manager [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 522.596789] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Acquiring lock "refresh_cache-3a9b16bb-9d58-4f0a-8fab-16f2486a6702" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 522.597328] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Acquired lock "refresh_cache-3a9b16bb-9d58-4f0a-8fab-16f2486a6702" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 522.597328] env[62569]: DEBUG nova.network.neutron [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 522.707915] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cc644a3-5ff2-49f9-ae71-eb2cfa83fc8b tempest-InstanceActionsV221TestJSON-879012145 tempest-InstanceActionsV221TestJSON-879012145-project-member] Lock "99412ace-3825-4c62-ad54-e59b83218387" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.508s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.798257] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.849411] env[62569]: DEBUG nova.compute.manager [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 522.882521] env[62569]: DEBUG nova.network.neutron [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 522.890536] env[62569]: DEBUG nova.virt.hardware [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 522.890536] env[62569]: DEBUG nova.virt.hardware [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 522.890536] env[62569]: DEBUG nova.virt.hardware [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 522.890876] env[62569]: DEBUG nova.virt.hardware [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 522.890876] env[62569]: DEBUG nova.virt.hardware [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 522.890876] env[62569]: DEBUG nova.virt.hardware [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 522.890876] env[62569]: DEBUG nova.virt.hardware [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 522.890876] env[62569]: DEBUG nova.virt.hardware [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 522.892017] env[62569]: DEBUG nova.virt.hardware [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 522.892017] env[62569]: DEBUG nova.virt.hardware [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 522.892017] env[62569]: DEBUG nova.virt.hardware [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 522.892241] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e93a528-3c4f-444e-9389-d3185aaf3f98 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.909637] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a0738bb-ce63-440b-bfce-9c9ec4e54078 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.024108] env[62569]: DEBUG nova.network.neutron [-] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 523.035222] env[62569]: ERROR nova.compute.manager [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 95c9ce89-4e44-45c7-8dcb-4c285a137ff2, please check neutron logs for more information. [ 523.035222] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 523.035222] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 523.035222] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 523.035222] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.035222] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 523.035222] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.035222] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 523.035222] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.035222] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 523.035222] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.035222] env[62569]: ERROR nova.compute.manager raise self.value [ 523.035222] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.035222] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 523.035222] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.035222] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 523.035773] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.035773] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 523.035773] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 95c9ce89-4e44-45c7-8dcb-4c285a137ff2, please check neutron logs for more information. [ 523.035773] env[62569]: ERROR nova.compute.manager [ 523.035773] env[62569]: Traceback (most recent call last): [ 523.035773] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 523.035773] env[62569]: listener.cb(fileno) [ 523.035773] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.035773] env[62569]: result = function(*args, **kwargs) [ 523.035773] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 523.035773] env[62569]: return func(*args, **kwargs) [ 523.035773] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 523.035773] env[62569]: raise e [ 523.035773] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 523.035773] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 523.035773] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.035773] env[62569]: created_port_ids = self._update_ports_for_instance( [ 523.035773] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.035773] env[62569]: with excutils.save_and_reraise_exception(): [ 523.035773] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.035773] env[62569]: self.force_reraise() [ 523.035773] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.035773] env[62569]: raise self.value [ 523.035773] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.035773] env[62569]: updated_port = self._update_port( [ 523.035773] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.035773] env[62569]: _ensure_no_port_binding_failure(port) [ 523.035773] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.035773] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 523.036673] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 95c9ce89-4e44-45c7-8dcb-4c285a137ff2, please check neutron logs for more information. [ 523.036673] env[62569]: Removing descriptor: 17 [ 523.036673] env[62569]: ERROR nova.compute.manager [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 95c9ce89-4e44-45c7-8dcb-4c285a137ff2, please check neutron logs for more information. [ 523.036673] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Traceback (most recent call last): [ 523.036673] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 523.036673] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] yield resources [ 523.036673] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 523.036673] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] self.driver.spawn(context, instance, image_meta, [ 523.036673] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 523.036673] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] self._vmops.spawn(context, instance, image_meta, injected_files, [ 523.036673] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 523.036673] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] vm_ref = self.build_virtual_machine(instance, [ 523.036988] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 523.036988] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] vif_infos = vmwarevif.get_vif_info(self._session, [ 523.036988] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 523.036988] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] for vif in network_info: [ 523.036988] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 523.036988] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] return self._sync_wrapper(fn, *args, **kwargs) [ 523.036988] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 523.036988] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] self.wait() [ 523.036988] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 523.036988] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] self[:] = self._gt.wait() [ 523.036988] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 523.036988] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] return self._exit_event.wait() [ 523.036988] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 523.037346] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] result = hub.switch() [ 523.037346] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 523.037346] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] return self.greenlet.switch() [ 523.037346] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.037346] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] result = function(*args, **kwargs) [ 523.037346] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 523.037346] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] return func(*args, **kwargs) [ 523.037346] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 523.037346] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] raise e [ 523.037346] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 523.037346] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] nwinfo = self.network_api.allocate_for_instance( [ 523.037346] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.037346] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] created_port_ids = self._update_ports_for_instance( [ 523.037775] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.037775] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] with excutils.save_and_reraise_exception(): [ 523.037775] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.037775] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] self.force_reraise() [ 523.037775] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.037775] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] raise self.value [ 523.037775] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.037775] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] updated_port = self._update_port( [ 523.037775] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.037775] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] _ensure_no_port_binding_failure(port) [ 523.037775] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.037775] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] raise exception.PortBindingFailed(port_id=port['id']) [ 523.038143] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] nova.exception.PortBindingFailed: Binding failed for port 95c9ce89-4e44-45c7-8dcb-4c285a137ff2, please check neutron logs for more information. [ 523.038143] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] [ 523.038143] env[62569]: INFO nova.compute.manager [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Terminating instance [ 523.125890] env[62569]: DEBUG nova.network.neutron [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 523.158100] env[62569]: DEBUG nova.network.neutron [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Successfully created port: 2077b280-333c-4da3-90fa-8a1a54c28e44 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 523.314377] env[62569]: DEBUG nova.network.neutron [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 523.335868] env[62569]: DEBUG nova.compute.manager [req-19c68e13-5072-4999-934e-08449a98c36f req-30586339-99ea-4aec-b69b-bc6ed26538c4 service nova] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Received event network-changed-409365e0-c0a3-4367-add1-3803636b9262 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 523.336137] env[62569]: DEBUG nova.compute.manager [req-19c68e13-5072-4999-934e-08449a98c36f req-30586339-99ea-4aec-b69b-bc6ed26538c4 service nova] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Refreshing instance network info cache due to event network-changed-409365e0-c0a3-4367-add1-3803636b9262. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 523.336306] env[62569]: DEBUG oslo_concurrency.lockutils [req-19c68e13-5072-4999-934e-08449a98c36f req-30586339-99ea-4aec-b69b-bc6ed26538c4 service nova] Acquiring lock "refresh_cache-ff90249f-ae4b-4aeb-91ed-d8a41a942a5d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.337098] env[62569]: DEBUG oslo_concurrency.lockutils [req-19c68e13-5072-4999-934e-08449a98c36f req-30586339-99ea-4aec-b69b-bc6ed26538c4 service nova] Acquired lock "refresh_cache-ff90249f-ae4b-4aeb-91ed-d8a41a942a5d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.337620] env[62569]: DEBUG nova.network.neutron [req-19c68e13-5072-4999-934e-08449a98c36f req-30586339-99ea-4aec-b69b-bc6ed26538c4 service nova] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Refreshing network info cache for port 409365e0-c0a3-4367-add1-3803636b9262 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 523.492682] env[62569]: DEBUG nova.network.neutron [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 523.525408] env[62569]: INFO nova.compute.manager [-] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Took 1.07 seconds to deallocate network for instance. [ 523.527837] env[62569]: DEBUG nova.compute.claims [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 523.528029] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.542084] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Acquiring lock "refresh_cache-832ec210-bb14-4a52-9634-50b7709a9976" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.542297] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Acquired lock "refresh_cache-832ec210-bb14-4a52-9634-50b7709a9976" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.542431] env[62569]: DEBUG nova.network.neutron [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 523.811688] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f3d80bd-87b8-4913-b5af-ddd7bfd035ab {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.820864] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Releasing lock "refresh_cache-3a9b16bb-9d58-4f0a-8fab-16f2486a6702" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 523.821147] env[62569]: DEBUG nova.compute.manager [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 523.821409] env[62569]: DEBUG nova.compute.manager [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 523.821496] env[62569]: DEBUG nova.network.neutron [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 523.824498] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfdc202b-e650-4b0d-a34f-1e77b59328ca {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.859814] env[62569]: DEBUG nova.network.neutron [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 523.861762] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d637209-d8cc-4bd9-ad79-d73f2d8733c9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.870229] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ea21cfe-a34e-42f2-a614-c138aa6ce062 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.884839] env[62569]: DEBUG nova.compute.provider_tree [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 523.886862] env[62569]: DEBUG nova.network.neutron [req-19c68e13-5072-4999-934e-08449a98c36f req-30586339-99ea-4aec-b69b-bc6ed26538c4 service nova] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 523.895152] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Acquiring lock "9fa91603-f812-4d93-bef1-58e455f2c2e8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.895152] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Lock "9fa91603-f812-4d93-bef1-58e455f2c2e8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.945837] env[62569]: DEBUG nova.network.neutron [req-19c68e13-5072-4999-934e-08449a98c36f req-30586339-99ea-4aec-b69b-bc6ed26538c4 service nova] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 523.994463] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Releasing lock "refresh_cache-17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 523.994463] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Updated the network info_cache for instance {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10379}} [ 523.994463] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 523.994738] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 523.994861] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 523.995063] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 523.995256] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 523.995437] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._sync_power_states {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 524.063575] env[62569]: DEBUG nova.network.neutron [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 524.118706] env[62569]: DEBUG nova.network.neutron [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.184696] env[62569]: DEBUG nova.compute.manager [None req-bb330c44-24fd-404e-a318-5d9e5efc2797 tempest-ServerDiagnosticsV248Test-587967259 tempest-ServerDiagnosticsV248Test-587967259-project-admin] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 524.185858] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3df1b41c-5e3c-471d-804a-d4c05b2bef82 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.193157] env[62569]: INFO nova.compute.manager [None req-bb330c44-24fd-404e-a318-5d9e5efc2797 tempest-ServerDiagnosticsV248Test-587967259 tempest-ServerDiagnosticsV248Test-587967259-project-admin] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Retrieving diagnostics [ 524.194155] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630a91b5-46a6-4f24-a6ca-03325c668c03 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.365443] env[62569]: DEBUG nova.network.neutron [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.394181] env[62569]: DEBUG nova.scheduler.client.report [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 524.401199] env[62569]: DEBUG nova.compute.manager [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 524.449221] env[62569]: DEBUG oslo_concurrency.lockutils [req-19c68e13-5072-4999-934e-08449a98c36f req-30586339-99ea-4aec-b69b-bc6ed26538c4 service nova] Releasing lock "refresh_cache-ff90249f-ae4b-4aeb-91ed-d8a41a942a5d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.499484] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Getting list of instances from cluster (obj){ [ 524.499484] env[62569]: value = "domain-c8" [ 524.499484] env[62569]: _type = "ClusterComputeResource" [ 524.499484] env[62569]: } {{(pid=62569) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 524.500626] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d77d6869-89b9-4263-8ac8-a3930f7b9a19 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.510227] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Got total of 1 instances {{(pid=62569) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 524.510380] env[62569]: WARNING nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] While synchronizing instance power states, found 6 instances in the database and 1 instances on the hypervisor. [ 524.510522] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Triggering sync for uuid 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007 {{(pid=62569) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10718}} [ 524.510705] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Triggering sync for uuid 611d4cfa-a888-47bc-8857-f061046a77ed {{(pid=62569) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10718}} [ 524.510859] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Triggering sync for uuid ff90249f-ae4b-4aeb-91ed-d8a41a942a5d {{(pid=62569) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10718}} [ 524.511015] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Triggering sync for uuid 832ec210-bb14-4a52-9634-50b7709a9976 {{(pid=62569) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10718}} [ 524.511166] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Triggering sync for uuid 597c9673-bbd4-4dd0-829a-c05f8667c951 {{(pid=62569) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10718}} [ 524.511498] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Triggering sync for uuid c542f5dd-10d8-4959-a3ab-eab9b8f4525a {{(pid=62569) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10718}} [ 524.511667] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.511783] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.512035] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "611d4cfa-a888-47bc-8857-f061046a77ed" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.512250] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "ff90249f-ae4b-4aeb-91ed-d8a41a942a5d" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.512445] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "832ec210-bb14-4a52-9634-50b7709a9976" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.512633] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "597c9673-bbd4-4dd0-829a-c05f8667c951" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.512835] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "c542f5dd-10d8-4959-a3ab-eab9b8f4525a" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.513010] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 524.513231] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62569) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 524.513894] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47012db6-c90b-4370-ba6f-277b43dcd23d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.516291] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 524.623798] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Releasing lock "refresh_cache-832ec210-bb14-4a52-9634-50b7709a9976" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.624242] env[62569]: DEBUG nova.compute.manager [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 524.624435] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 524.624849] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f50a8e87-2648-4da9-ac51-4248da79eb1f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.633630] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6abfdd97-ea3a-4871-8c9c-f6f56054daa5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.655220] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 832ec210-bb14-4a52-9634-50b7709a9976 could not be found. [ 524.655442] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 524.656095] env[62569]: INFO nova.compute.manager [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Took 0.03 seconds to destroy the instance on the hypervisor. [ 524.656095] env[62569]: DEBUG oslo.service.loopingcall [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 524.656095] env[62569]: DEBUG nova.compute.manager [-] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 524.656579] env[62569]: DEBUG nova.network.neutron [-] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 524.672300] env[62569]: DEBUG nova.network.neutron [-] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 524.869843] env[62569]: INFO nova.compute.manager [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] [instance: 3a9b16bb-9d58-4f0a-8fab-16f2486a6702] Took 1.05 seconds to deallocate network for instance. [ 524.918824] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.334s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 524.919513] env[62569]: DEBUG nova.compute.manager [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 524.927035] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.695s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.964373] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.019433] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.024053] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.512s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 525.177621] env[62569]: DEBUG nova.network.neutron [-] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.432941] env[62569]: DEBUG nova.compute.utils [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 525.440021] env[62569]: DEBUG nova.compute.manager [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 525.440021] env[62569]: DEBUG nova.network.neutron [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 525.532709] env[62569]: DEBUG nova.policy [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '626ce48f369d40df9fb2ffd437b6e6d2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2a4472f7f47461e844405b5cabe23eb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 525.619529] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb6ff65-84c5-4760-8bc0-1b7bc1d54a9b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.632154] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d556a63-f60f-462c-b038-75133aa8c8ab {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.669117] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66f05ec9-05bd-475f-bf69-56855f5f3266 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.681341] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bebe13d-4720-4a68-b4c8-dc806d230d6b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.685731] env[62569]: INFO nova.compute.manager [-] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Took 1.03 seconds to deallocate network for instance. [ 525.699239] env[62569]: DEBUG nova.compute.provider_tree [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 525.699239] env[62569]: DEBUG nova.compute.claims [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 525.699424] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.908303] env[62569]: INFO nova.scheduler.client.report [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Deleted allocations for instance 3a9b16bb-9d58-4f0a-8fab-16f2486a6702 [ 525.939723] env[62569]: DEBUG nova.compute.manager [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 525.944203] env[62569]: DEBUG nova.compute.manager [req-c4defbbb-c68e-46b3-bf1e-76f8f6db5f1a req-f1377a61-82c8-401d-8882-5e5d6d78928b service nova] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Received event network-changed-95c9ce89-4e44-45c7-8dcb-4c285a137ff2 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 525.946698] env[62569]: DEBUG nova.compute.manager [req-c4defbbb-c68e-46b3-bf1e-76f8f6db5f1a req-f1377a61-82c8-401d-8882-5e5d6d78928b service nova] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Refreshing instance network info cache due to event network-changed-95c9ce89-4e44-45c7-8dcb-4c285a137ff2. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 525.946698] env[62569]: DEBUG oslo_concurrency.lockutils [req-c4defbbb-c68e-46b3-bf1e-76f8f6db5f1a req-f1377a61-82c8-401d-8882-5e5d6d78928b service nova] Acquiring lock "refresh_cache-832ec210-bb14-4a52-9634-50b7709a9976" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 525.946698] env[62569]: DEBUG oslo_concurrency.lockutils [req-c4defbbb-c68e-46b3-bf1e-76f8f6db5f1a req-f1377a61-82c8-401d-8882-5e5d6d78928b service nova] Acquired lock "refresh_cache-832ec210-bb14-4a52-9634-50b7709a9976" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 525.946698] env[62569]: DEBUG nova.network.neutron [req-c4defbbb-c68e-46b3-bf1e-76f8f6db5f1a req-f1377a61-82c8-401d-8882-5e5d6d78928b service nova] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Refreshing network info cache for port 95c9ce89-4e44-45c7-8dcb-4c285a137ff2 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 526.126921] env[62569]: ERROR nova.compute.manager [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2077b280-333c-4da3-90fa-8a1a54c28e44, please check neutron logs for more information. [ 526.126921] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 526.126921] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 526.126921] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 526.126921] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 526.126921] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 526.126921] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 526.126921] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 526.126921] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 526.126921] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 526.126921] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 526.126921] env[62569]: ERROR nova.compute.manager raise self.value [ 526.126921] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 526.126921] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 526.126921] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 526.126921] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 526.127386] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 526.127386] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 526.127386] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2077b280-333c-4da3-90fa-8a1a54c28e44, please check neutron logs for more information. [ 526.127386] env[62569]: ERROR nova.compute.manager [ 526.127386] env[62569]: Traceback (most recent call last): [ 526.127386] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 526.127386] env[62569]: listener.cb(fileno) [ 526.127386] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 526.127386] env[62569]: result = function(*args, **kwargs) [ 526.127386] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 526.127386] env[62569]: return func(*args, **kwargs) [ 526.127386] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 526.127386] env[62569]: raise e [ 526.127386] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 526.127386] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 526.127386] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 526.127386] env[62569]: created_port_ids = self._update_ports_for_instance( [ 526.127386] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 526.127386] env[62569]: with excutils.save_and_reraise_exception(): [ 526.127386] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 526.127386] env[62569]: self.force_reraise() [ 526.127386] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 526.127386] env[62569]: raise self.value [ 526.127386] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 526.127386] env[62569]: updated_port = self._update_port( [ 526.127386] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 526.127386] env[62569]: _ensure_no_port_binding_failure(port) [ 526.127386] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 526.127386] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 526.128602] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 2077b280-333c-4da3-90fa-8a1a54c28e44, please check neutron logs for more information. [ 526.128602] env[62569]: Removing descriptor: 16 [ 526.128602] env[62569]: ERROR nova.compute.manager [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2077b280-333c-4da3-90fa-8a1a54c28e44, please check neutron logs for more information. [ 526.128602] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Traceback (most recent call last): [ 526.128602] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 526.128602] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] yield resources [ 526.128602] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 526.128602] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] self.driver.spawn(context, instance, image_meta, [ 526.128602] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 526.128602] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] self._vmops.spawn(context, instance, image_meta, injected_files, [ 526.128602] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 526.128602] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] vm_ref = self.build_virtual_machine(instance, [ 526.128946] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 526.128946] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] vif_infos = vmwarevif.get_vif_info(self._session, [ 526.128946] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 526.128946] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] for vif in network_info: [ 526.128946] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 526.128946] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] return self._sync_wrapper(fn, *args, **kwargs) [ 526.128946] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 526.128946] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] self.wait() [ 526.128946] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 526.128946] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] self[:] = self._gt.wait() [ 526.128946] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 526.128946] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] return self._exit_event.wait() [ 526.128946] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 526.129332] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] result = hub.switch() [ 526.129332] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 526.129332] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] return self.greenlet.switch() [ 526.129332] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 526.129332] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] result = function(*args, **kwargs) [ 526.129332] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 526.129332] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] return func(*args, **kwargs) [ 526.129332] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 526.129332] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] raise e [ 526.129332] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 526.129332] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] nwinfo = self.network_api.allocate_for_instance( [ 526.129332] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 526.129332] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] created_port_ids = self._update_ports_for_instance( [ 526.129696] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 526.129696] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] with excutils.save_and_reraise_exception(): [ 526.129696] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 526.129696] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] self.force_reraise() [ 526.129696] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 526.129696] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] raise self.value [ 526.129696] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 526.129696] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] updated_port = self._update_port( [ 526.129696] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 526.129696] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] _ensure_no_port_binding_failure(port) [ 526.129696] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 526.129696] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] raise exception.PortBindingFailed(port_id=port['id']) [ 526.130085] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] nova.exception.PortBindingFailed: Binding failed for port 2077b280-333c-4da3-90fa-8a1a54c28e44, please check neutron logs for more information. [ 526.130085] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] [ 526.130085] env[62569]: INFO nova.compute.manager [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Terminating instance [ 526.202941] env[62569]: DEBUG nova.scheduler.client.report [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 526.242148] env[62569]: DEBUG nova.network.neutron [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Successfully created port: 247970bb-e8a0-4e10-a07a-04e6c4ced856 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 526.339092] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Acquiring lock "5b1117bd-6210-4df2-b828-f76951752f2f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.339420] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Lock "5b1117bd-6210-4df2-b828-f76951752f2f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.421409] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc24652a-c450-4b4a-bda6-1b99c3598387 tempest-ImagesOneServerTestJSON-631823963 tempest-ImagesOneServerTestJSON-631823963-project-member] Lock "3a9b16bb-9d58-4f0a-8fab-16f2486a6702" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.366s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 526.490015] env[62569]: DEBUG nova.network.neutron [req-c4defbbb-c68e-46b3-bf1e-76f8f6db5f1a req-f1377a61-82c8-401d-8882-5e5d6d78928b service nova] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 526.588047] env[62569]: DEBUG nova.network.neutron [req-c4defbbb-c68e-46b3-bf1e-76f8f6db5f1a req-f1377a61-82c8-401d-8882-5e5d6d78928b service nova] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.634513] env[62569]: DEBUG oslo_concurrency.lockutils [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Acquiring lock "refresh_cache-597c9673-bbd4-4dd0-829a-c05f8667c951" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 526.635040] env[62569]: DEBUG oslo_concurrency.lockutils [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Acquired lock "refresh_cache-597c9673-bbd4-4dd0-829a-c05f8667c951" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 526.636960] env[62569]: DEBUG nova.network.neutron [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 526.709445] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.782s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 526.709860] env[62569]: ERROR nova.compute.manager [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8e17f59f-041c-407b-bdfa-f054585d93d7, please check neutron logs for more information. [ 526.709860] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Traceback (most recent call last): [ 526.709860] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 526.709860] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] self.driver.spawn(context, instance, image_meta, [ 526.709860] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 526.709860] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] self._vmops.spawn(context, instance, image_meta, injected_files, [ 526.709860] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 526.709860] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] vm_ref = self.build_virtual_machine(instance, [ 526.709860] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 526.709860] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] vif_infos = vmwarevif.get_vif_info(self._session, [ 526.709860] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 526.710182] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] for vif in network_info: [ 526.710182] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 526.710182] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] return self._sync_wrapper(fn, *args, **kwargs) [ 526.710182] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 526.710182] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] self.wait() [ 526.710182] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 526.710182] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] self[:] = self._gt.wait() [ 526.710182] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 526.710182] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] return self._exit_event.wait() [ 526.710182] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 526.710182] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] current.throw(*self._exc) [ 526.710182] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 526.710182] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] result = function(*args, **kwargs) [ 526.710674] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 526.710674] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] return func(*args, **kwargs) [ 526.710674] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 526.710674] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] raise e [ 526.710674] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 526.710674] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] nwinfo = self.network_api.allocate_for_instance( [ 526.710674] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 526.710674] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] created_port_ids = self._update_ports_for_instance( [ 526.710674] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 526.710674] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] with excutils.save_and_reraise_exception(): [ 526.710674] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 526.710674] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] self.force_reraise() [ 526.710674] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 526.710999] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] raise self.value [ 526.710999] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 526.710999] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] updated_port = self._update_port( [ 526.710999] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 526.710999] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] _ensure_no_port_binding_failure(port) [ 526.710999] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 526.710999] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] raise exception.PortBindingFailed(port_id=port['id']) [ 526.710999] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] nova.exception.PortBindingFailed: Binding failed for port 8e17f59f-041c-407b-bdfa-f054585d93d7, please check neutron logs for more information. [ 526.710999] env[62569]: ERROR nova.compute.manager [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] [ 526.710999] env[62569]: DEBUG nova.compute.utils [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Binding failed for port 8e17f59f-041c-407b-bdfa-f054585d93d7, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 526.714403] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.886s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.715972] env[62569]: INFO nova.compute.claims [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 526.724794] env[62569]: DEBUG nova.compute.manager [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Build of instance 611d4cfa-a888-47bc-8857-f061046a77ed was re-scheduled: Binding failed for port 8e17f59f-041c-407b-bdfa-f054585d93d7, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 526.725361] env[62569]: DEBUG nova.compute.manager [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 526.725618] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Acquiring lock "refresh_cache-611d4cfa-a888-47bc-8857-f061046a77ed" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 526.725780] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Acquired lock "refresh_cache-611d4cfa-a888-47bc-8857-f061046a77ed" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 526.725954] env[62569]: DEBUG nova.network.neutron [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 526.925290] env[62569]: DEBUG nova.compute.manager [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 526.957011] env[62569]: DEBUG nova.compute.manager [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 526.994607] env[62569]: DEBUG nova.virt.hardware [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:31:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='2073256351',id=22,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-962625974',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 526.994959] env[62569]: DEBUG nova.virt.hardware [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 526.995463] env[62569]: DEBUG nova.virt.hardware [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 526.995774] env[62569]: DEBUG nova.virt.hardware [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 526.996013] env[62569]: DEBUG nova.virt.hardware [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 526.996381] env[62569]: DEBUG nova.virt.hardware [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 527.001128] env[62569]: DEBUG nova.virt.hardware [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 527.001128] env[62569]: DEBUG nova.virt.hardware [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 527.001128] env[62569]: DEBUG nova.virt.hardware [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 527.001128] env[62569]: DEBUG nova.virt.hardware [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 527.001128] env[62569]: DEBUG nova.virt.hardware [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 527.001777] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40d9bfe5-b366-491d-b9f8-f5835126a539 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.012875] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b5c0cea-4a42-4027-9ca4-ffd78c8d5bf0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.090634] env[62569]: DEBUG oslo_concurrency.lockutils [req-c4defbbb-c68e-46b3-bf1e-76f8f6db5f1a req-f1377a61-82c8-401d-8882-5e5d6d78928b service nova] Releasing lock "refresh_cache-832ec210-bb14-4a52-9634-50b7709a9976" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.188939] env[62569]: DEBUG nova.network.neutron [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 527.212577] env[62569]: DEBUG nova.compute.manager [req-6bac400e-48b6-4f2c-82ac-fd5c7c8b1205 req-11ced8da-500a-4382-bbb5-5ce089cfa670 service nova] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Received event network-vif-deleted-409365e0-c0a3-4367-add1-3803636b9262 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 527.223281] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquiring lock "232a126f-37ec-4aea-a602-eb3a13d1d70a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.224028] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Lock "232a126f-37ec-4aea-a602-eb3a13d1d70a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.257712] env[62569]: DEBUG nova.network.neutron [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 527.274799] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquiring lock "11c53593-cf14-4e46-b546-c90a3184028e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.275036] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Lock "11c53593-cf14-4e46-b546-c90a3184028e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.301804] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquiring lock "a9224414-6d20-4d5f-be5d-2b893117f4b8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.302820] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Lock "a9224414-6d20-4d5f-be5d-2b893117f4b8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.381190] env[62569]: DEBUG nova.network.neutron [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.452895] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.457740] env[62569]: DEBUG nova.network.neutron [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.890660] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Releasing lock "refresh_cache-611d4cfa-a888-47bc-8857-f061046a77ed" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.890716] env[62569]: DEBUG nova.compute.manager [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 527.891635] env[62569]: DEBUG nova.compute.manager [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 527.891635] env[62569]: DEBUG nova.network.neutron [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 527.919030] env[62569]: DEBUG nova.network.neutron [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 527.920989] env[62569]: DEBUG oslo_concurrency.lockutils [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Acquiring lock "49fdf371-5278-442b-b8cc-9a0fc22947b4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.920989] env[62569]: DEBUG oslo_concurrency.lockutils [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Lock "49fdf371-5278-442b-b8cc-9a0fc22947b4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.963758] env[62569]: DEBUG oslo_concurrency.lockutils [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Releasing lock "refresh_cache-597c9673-bbd4-4dd0-829a-c05f8667c951" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.964080] env[62569]: DEBUG nova.compute.manager [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 527.964263] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 527.965410] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ec638039-02c6-4fea-b1fe-da1b6fdc0b99 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.974060] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-633579d1-1185-4aa9-b8af-fc152006a792 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.002197] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 597c9673-bbd4-4dd0-829a-c05f8667c951 could not be found. [ 528.003020] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 528.003020] env[62569]: INFO nova.compute.manager [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Took 0.04 seconds to destroy the instance on the hypervisor. [ 528.003170] env[62569]: DEBUG oslo.service.loopingcall [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 528.006301] env[62569]: DEBUG nova.compute.manager [-] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 528.006504] env[62569]: DEBUG nova.network.neutron [-] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 528.020379] env[62569]: ERROR nova.compute.manager [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 247970bb-e8a0-4e10-a07a-04e6c4ced856, please check neutron logs for more information. [ 528.020379] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 528.020379] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 528.020379] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 528.020379] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 528.020379] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 528.020379] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 528.020379] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 528.020379] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 528.020379] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 528.020379] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 528.020379] env[62569]: ERROR nova.compute.manager raise self.value [ 528.020379] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 528.020379] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 528.020379] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 528.020379] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 528.020852] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 528.020852] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 528.020852] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 247970bb-e8a0-4e10-a07a-04e6c4ced856, please check neutron logs for more information. [ 528.020852] env[62569]: ERROR nova.compute.manager [ 528.020852] env[62569]: Traceback (most recent call last): [ 528.020852] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 528.020852] env[62569]: listener.cb(fileno) [ 528.020852] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 528.020852] env[62569]: result = function(*args, **kwargs) [ 528.020852] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 528.020852] env[62569]: return func(*args, **kwargs) [ 528.020852] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 528.020852] env[62569]: raise e [ 528.020852] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 528.020852] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 528.020852] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 528.020852] env[62569]: created_port_ids = self._update_ports_for_instance( [ 528.020852] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 528.020852] env[62569]: with excutils.save_and_reraise_exception(): [ 528.020852] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 528.020852] env[62569]: self.force_reraise() [ 528.020852] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 528.020852] env[62569]: raise self.value [ 528.020852] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 528.020852] env[62569]: updated_port = self._update_port( [ 528.020852] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 528.020852] env[62569]: _ensure_no_port_binding_failure(port) [ 528.020852] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 528.020852] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 528.021614] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 247970bb-e8a0-4e10-a07a-04e6c4ced856, please check neutron logs for more information. [ 528.021614] env[62569]: Removing descriptor: 14 [ 528.021614] env[62569]: ERROR nova.compute.manager [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 247970bb-e8a0-4e10-a07a-04e6c4ced856, please check neutron logs for more information. [ 528.021614] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Traceback (most recent call last): [ 528.021614] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 528.021614] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] yield resources [ 528.021614] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 528.021614] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] self.driver.spawn(context, instance, image_meta, [ 528.021614] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 528.021614] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 528.021614] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 528.021614] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] vm_ref = self.build_virtual_machine(instance, [ 528.022011] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 528.022011] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] vif_infos = vmwarevif.get_vif_info(self._session, [ 528.022011] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 528.022011] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] for vif in network_info: [ 528.022011] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 528.022011] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] return self._sync_wrapper(fn, *args, **kwargs) [ 528.022011] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 528.022011] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] self.wait() [ 528.022011] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 528.022011] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] self[:] = self._gt.wait() [ 528.022011] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 528.022011] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] return self._exit_event.wait() [ 528.022011] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 528.022371] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] result = hub.switch() [ 528.022371] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 528.022371] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] return self.greenlet.switch() [ 528.022371] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 528.022371] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] result = function(*args, **kwargs) [ 528.022371] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 528.022371] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] return func(*args, **kwargs) [ 528.022371] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 528.022371] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] raise e [ 528.022371] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 528.022371] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] nwinfo = self.network_api.allocate_for_instance( [ 528.022371] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 528.022371] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] created_port_ids = self._update_ports_for_instance( [ 528.022698] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 528.022698] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] with excutils.save_and_reraise_exception(): [ 528.022698] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 528.022698] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] self.force_reraise() [ 528.022698] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 528.022698] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] raise self.value [ 528.022698] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 528.022698] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] updated_port = self._update_port( [ 528.022698] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 528.022698] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] _ensure_no_port_binding_failure(port) [ 528.022698] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 528.022698] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] raise exception.PortBindingFailed(port_id=port['id']) [ 528.023112] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] nova.exception.PortBindingFailed: Binding failed for port 247970bb-e8a0-4e10-a07a-04e6c4ced856, please check neutron logs for more information. [ 528.023112] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] [ 528.023112] env[62569]: INFO nova.compute.manager [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Terminating instance [ 528.027292] env[62569]: DEBUG nova.network.neutron [-] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 528.074325] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a717f727-26b0-42dd-90c7-cd68fe780d2f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.086731] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8db0a97d-b673-4c0b-adcb-80f4da820a2f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.131769] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-104c8002-ad69-45e2-a41d-e9423434dd5b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.141694] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6f6699c-2d09-4bec-b15f-22b115070c99 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.161160] env[62569]: DEBUG nova.compute.provider_tree [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 528.422350] env[62569]: DEBUG nova.network.neutron [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.535219] env[62569]: DEBUG nova.network.neutron [-] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.535219] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Acquiring lock "refresh_cache-c542f5dd-10d8-4959-a3ab-eab9b8f4525a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 528.535219] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Acquired lock "refresh_cache-c542f5dd-10d8-4959-a3ab-eab9b8f4525a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 528.535219] env[62569]: DEBUG nova.network.neutron [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 528.666815] env[62569]: DEBUG nova.scheduler.client.report [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 528.926306] env[62569]: INFO nova.compute.manager [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] Took 1.03 seconds to deallocate network for instance. [ 529.039932] env[62569]: INFO nova.compute.manager [-] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Took 1.03 seconds to deallocate network for instance. [ 529.045726] env[62569]: DEBUG nova.compute.claims [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 529.045851] env[62569]: DEBUG oslo_concurrency.lockutils [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.066250] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Acquiring lock "ef52ac4a-fa98-4af8-b08a-b34cae526850" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.066346] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Lock "ef52ac4a-fa98-4af8-b08a-b34cae526850" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.103627] env[62569]: DEBUG nova.network.neutron [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 529.174808] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.460s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.175350] env[62569]: DEBUG nova.compute.manager [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 529.181439] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.112s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.186433] env[62569]: INFO nova.compute.claims [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 529.312105] env[62569]: DEBUG nova.network.neutron [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.486106] env[62569]: DEBUG nova.compute.manager [req-1a803d41-b287-4fd4-84cc-4c9149eb54b6 req-5afdba9e-46dc-4121-98a6-1e430eb318fe service nova] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Received event network-vif-deleted-95c9ce89-4e44-45c7-8dcb-4c285a137ff2 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 529.487024] env[62569]: DEBUG nova.compute.manager [req-1a803d41-b287-4fd4-84cc-4c9149eb54b6 req-5afdba9e-46dc-4121-98a6-1e430eb318fe service nova] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Received event network-changed-247970bb-e8a0-4e10-a07a-04e6c4ced856 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 529.487024] env[62569]: DEBUG nova.compute.manager [req-1a803d41-b287-4fd4-84cc-4c9149eb54b6 req-5afdba9e-46dc-4121-98a6-1e430eb318fe service nova] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Refreshing instance network info cache due to event network-changed-247970bb-e8a0-4e10-a07a-04e6c4ced856. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 529.487024] env[62569]: DEBUG oslo_concurrency.lockutils [req-1a803d41-b287-4fd4-84cc-4c9149eb54b6 req-5afdba9e-46dc-4121-98a6-1e430eb318fe service nova] Acquiring lock "refresh_cache-c542f5dd-10d8-4959-a3ab-eab9b8f4525a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 529.697253] env[62569]: DEBUG nova.compute.utils [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 529.706120] env[62569]: DEBUG nova.compute.manager [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 529.706120] env[62569]: DEBUG nova.network.neutron [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 529.820474] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Releasing lock "refresh_cache-c542f5dd-10d8-4959-a3ab-eab9b8f4525a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 529.820474] env[62569]: DEBUG nova.compute.manager [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 529.820474] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 529.820474] env[62569]: DEBUG oslo_concurrency.lockutils [req-1a803d41-b287-4fd4-84cc-4c9149eb54b6 req-5afdba9e-46dc-4121-98a6-1e430eb318fe service nova] Acquired lock "refresh_cache-c542f5dd-10d8-4959-a3ab-eab9b8f4525a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 529.820474] env[62569]: DEBUG nova.network.neutron [req-1a803d41-b287-4fd4-84cc-4c9149eb54b6 req-5afdba9e-46dc-4121-98a6-1e430eb318fe service nova] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Refreshing network info cache for port 247970bb-e8a0-4e10-a07a-04e6c4ced856 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 529.820956] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-58b4abc0-4e94-4b4f-bf86-09bbab69e39e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.833755] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2ddf2fb-6d95-4714-9d50-90e5e12c6f9d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.868190] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c542f5dd-10d8-4959-a3ab-eab9b8f4525a could not be found. [ 529.869381] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 529.869780] env[62569]: INFO nova.compute.manager [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Took 0.05 seconds to destroy the instance on the hypervisor. [ 529.870974] env[62569]: DEBUG oslo.service.loopingcall [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 529.871421] env[62569]: DEBUG nova.compute.manager [-] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 529.871672] env[62569]: DEBUG nova.network.neutron [-] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 529.936247] env[62569]: DEBUG nova.network.neutron [-] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 529.966509] env[62569]: INFO nova.scheduler.client.report [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Deleted allocations for instance 611d4cfa-a888-47bc-8857-f061046a77ed [ 530.110274] env[62569]: DEBUG nova.policy [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '57a76e8ca4ac418f9cebf177be57fd9e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cb955b265c874e86b329e0bb311e3094', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 530.207349] env[62569]: DEBUG nova.compute.manager [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 530.212740] env[62569]: DEBUG nova.compute.manager [req-a14f3427-1e27-407e-b9a5-e051f4746961 req-d43bfc6a-b949-49c9-992f-1a528b1f036e service nova] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Received event network-changed-2077b280-333c-4da3-90fa-8a1a54c28e44 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 530.212740] env[62569]: DEBUG nova.compute.manager [req-a14f3427-1e27-407e-b9a5-e051f4746961 req-d43bfc6a-b949-49c9-992f-1a528b1f036e service nova] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Refreshing instance network info cache due to event network-changed-2077b280-333c-4da3-90fa-8a1a54c28e44. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 530.212740] env[62569]: DEBUG oslo_concurrency.lockutils [req-a14f3427-1e27-407e-b9a5-e051f4746961 req-d43bfc6a-b949-49c9-992f-1a528b1f036e service nova] Acquiring lock "refresh_cache-597c9673-bbd4-4dd0-829a-c05f8667c951" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 530.212740] env[62569]: DEBUG oslo_concurrency.lockutils [req-a14f3427-1e27-407e-b9a5-e051f4746961 req-d43bfc6a-b949-49c9-992f-1a528b1f036e service nova] Acquired lock "refresh_cache-597c9673-bbd4-4dd0-829a-c05f8667c951" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 530.212920] env[62569]: DEBUG nova.network.neutron [req-a14f3427-1e27-407e-b9a5-e051f4746961 req-d43bfc6a-b949-49c9-992f-1a528b1f036e service nova] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Refreshing network info cache for port 2077b280-333c-4da3-90fa-8a1a54c28e44 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 530.340346] env[62569]: DEBUG nova.network.neutron [req-1a803d41-b287-4fd4-84cc-4c9149eb54b6 req-5afdba9e-46dc-4121-98a6-1e430eb318fe service nova] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 530.444016] env[62569]: DEBUG nova.network.neutron [-] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.456388] env[62569]: DEBUG oslo_concurrency.lockutils [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Acquiring lock "0dc705ca-6fd1-4844-9c2d-34f2c099ea67" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.456388] env[62569]: DEBUG oslo_concurrency.lockutils [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Lock "0dc705ca-6fd1-4844-9c2d-34f2c099ea67" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.479433] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9e8ff0b1-e626-4edc-9ad8-60762e4fa141 tempest-ServersAdminNegativeTestJSON-437662763 tempest-ServersAdminNegativeTestJSON-437662763-project-member] Lock "611d4cfa-a888-47bc-8857-f061046a77ed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.666s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 530.481169] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "611d4cfa-a888-47bc-8857-f061046a77ed" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 5.969s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.481169] env[62569]: INFO nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 611d4cfa-a888-47bc-8857-f061046a77ed] During sync_power_state the instance has a pending task (spawning). Skip. [ 530.481169] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "611d4cfa-a888-47bc-8857-f061046a77ed" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 530.531483] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-203a0f6a-4fec-4403-9bdf-2e8e4c0aea58 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.540689] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70570764-7b82-4f5f-b438-057c51891ce7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.545388] env[62569]: DEBUG nova.network.neutron [req-1a803d41-b287-4fd4-84cc-4c9149eb54b6 req-5afdba9e-46dc-4121-98a6-1e430eb318fe service nova] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.588029] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6baaeafc-2acc-424f-87c3-248b58e2ed71 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.595073] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d81b5421-4706-4ef5-93ed-f850de2f4c4d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.610063] env[62569]: DEBUG nova.compute.provider_tree [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 530.754864] env[62569]: DEBUG nova.network.neutron [req-a14f3427-1e27-407e-b9a5-e051f4746961 req-d43bfc6a-b949-49c9-992f-1a528b1f036e service nova] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 530.952536] env[62569]: INFO nova.compute.manager [-] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Took 1.08 seconds to deallocate network for instance. [ 530.955426] env[62569]: DEBUG nova.compute.claims [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 530.955709] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.986104] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 531.023333] env[62569]: DEBUG nova.network.neutron [req-a14f3427-1e27-407e-b9a5-e051f4746961 req-d43bfc6a-b949-49c9-992f-1a528b1f036e service nova] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.050559] env[62569]: DEBUG oslo_concurrency.lockutils [req-1a803d41-b287-4fd4-84cc-4c9149eb54b6 req-5afdba9e-46dc-4121-98a6-1e430eb318fe service nova] Releasing lock "refresh_cache-c542f5dd-10d8-4959-a3ab-eab9b8f4525a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.115425] env[62569]: DEBUG nova.scheduler.client.report [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 531.224437] env[62569]: DEBUG nova.compute.manager [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 531.262176] env[62569]: DEBUG nova.virt.hardware [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 531.262444] env[62569]: DEBUG nova.virt.hardware [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 531.262639] env[62569]: DEBUG nova.virt.hardware [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 531.262874] env[62569]: DEBUG nova.virt.hardware [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 531.263156] env[62569]: DEBUG nova.virt.hardware [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 531.263375] env[62569]: DEBUG nova.virt.hardware [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 531.263643] env[62569]: DEBUG nova.virt.hardware [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 531.263840] env[62569]: DEBUG nova.virt.hardware [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 531.264066] env[62569]: DEBUG nova.virt.hardware [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 531.264261] env[62569]: DEBUG nova.virt.hardware [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 531.264484] env[62569]: DEBUG nova.virt.hardware [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 531.265409] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d1ba2d6-2458-4302-b2d2-15e71ab3b1fa {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.275342] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14822071-bade-4c06-b5af-e1ac297cab39 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.514699] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.526556] env[62569]: DEBUG oslo_concurrency.lockutils [req-a14f3427-1e27-407e-b9a5-e051f4746961 req-d43bfc6a-b949-49c9-992f-1a528b1f036e service nova] Releasing lock "refresh_cache-597c9673-bbd4-4dd0-829a-c05f8667c951" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.526839] env[62569]: DEBUG nova.compute.manager [req-a14f3427-1e27-407e-b9a5-e051f4746961 req-d43bfc6a-b949-49c9-992f-1a528b1f036e service nova] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Received event network-vif-deleted-2077b280-333c-4da3-90fa-8a1a54c28e44 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 531.624917] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.444s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.624917] env[62569]: DEBUG nova.compute.manager [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 531.627698] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.830s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.629055] env[62569]: INFO nova.compute.claims [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 531.853142] env[62569]: DEBUG nova.network.neutron [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Successfully created port: 28017eb1-a7ac-4cfb-a089-ad527f6ed915 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 531.971468] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Acquiring lock "21af21da-87b1-4a84-8544-f1a88494eb08" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.971696] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Lock "21af21da-87b1-4a84-8544-f1a88494eb08" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.136204] env[62569]: DEBUG nova.compute.utils [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 532.137891] env[62569]: DEBUG nova.compute.manager [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 532.138226] env[62569]: DEBUG nova.network.neutron [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 532.284966] env[62569]: DEBUG nova.policy [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4b971531d3ff428cbea1742c1629d8ca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c917db0ac9124b9aa2be1e0a0bffb662', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 532.639292] env[62569]: DEBUG nova.compute.manager [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 532.967176] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fde6139f-065f-4e7a-8f80-2a3786e496d6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.979511] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88b464b6-55cc-413c-81d4-d5eaef6f2e6a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.016166] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdebd165-0621-4135-b759-25727cfdbc02 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.023475] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e3a2f27-f837-4d97-a28d-7dbef47e041e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.029849] env[62569]: DEBUG nova.compute.manager [req-282ffe6c-c211-4d8a-aa66-732cf8dfa9c8 req-4bd8cb94-3f1b-4822-a7ca-4840afa207f3 service nova] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Received event network-vif-deleted-247970bb-e8a0-4e10-a07a-04e6c4ced856 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 533.042961] env[62569]: DEBUG nova.compute.provider_tree [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 533.270579] env[62569]: DEBUG nova.network.neutron [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Successfully created port: ae242046-3aec-42af-a3cc-1839937c8497 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 533.547161] env[62569]: DEBUG nova.scheduler.client.report [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 533.658844] env[62569]: DEBUG nova.compute.manager [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 533.684191] env[62569]: DEBUG nova.virt.hardware [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 533.684309] env[62569]: DEBUG nova.virt.hardware [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 533.684526] env[62569]: DEBUG nova.virt.hardware [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 533.684747] env[62569]: DEBUG nova.virt.hardware [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 533.684951] env[62569]: DEBUG nova.virt.hardware [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 533.685386] env[62569]: DEBUG nova.virt.hardware [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 533.685648] env[62569]: DEBUG nova.virt.hardware [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 533.685827] env[62569]: DEBUG nova.virt.hardware [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 533.686108] env[62569]: DEBUG nova.virt.hardware [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 533.686198] env[62569]: DEBUG nova.virt.hardware [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 533.686622] env[62569]: DEBUG nova.virt.hardware [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 533.687692] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb733b3-fd70-4923-b700-f201d829ceeb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.701510] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0856ecf2-88ea-4886-aed1-153029a96934 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.055638] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.426s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 534.055638] env[62569]: DEBUG nova.compute.manager [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 534.057690] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.529s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.491903] env[62569]: DEBUG oslo_concurrency.lockutils [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Acquiring lock "c6292246-847f-4466-909b-6139401943af" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.493454] env[62569]: DEBUG oslo_concurrency.lockutils [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Lock "c6292246-847f-4466-909b-6139401943af" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.562908] env[62569]: DEBUG nova.compute.utils [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 534.567841] env[62569]: DEBUG nova.compute.manager [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Not allocating networking since 'none' was specified. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 534.974147] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7a4778a-d665-43a5-8bad-66407d6c531f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.982206] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dabccea9-fa63-4e88-86f6-6b5009e4dfe9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.019753] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd032256-d9fd-4246-beb3-a3a18a995e15 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.027580] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfe91891-50ac-4137-a013-3c0efddde765 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.042498] env[62569]: DEBUG nova.compute.provider_tree [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 535.068488] env[62569]: DEBUG nova.compute.manager [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 535.545769] env[62569]: DEBUG nova.scheduler.client.report [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 535.741711] env[62569]: DEBUG nova.compute.manager [None req-e573a96f-018e-4daf-8147-a59200f1993f tempest-ServerDiagnosticsV248Test-587967259 tempest-ServerDiagnosticsV248Test-587967259-project-admin] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 535.744024] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9c76349-fa2f-4e5b-9e0e-34b065196a1f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.753865] env[62569]: INFO nova.compute.manager [None req-e573a96f-018e-4daf-8147-a59200f1993f tempest-ServerDiagnosticsV248Test-587967259 tempest-ServerDiagnosticsV248Test-587967259-project-admin] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Retrieving diagnostics [ 535.754368] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5400183a-9823-4cf8-b773-b488aba9ba75 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.054606] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.997s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 536.054989] env[62569]: ERROR nova.compute.manager [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 409365e0-c0a3-4367-add1-3803636b9262, please check neutron logs for more information. [ 536.054989] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Traceback (most recent call last): [ 536.054989] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 536.054989] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] self.driver.spawn(context, instance, image_meta, [ 536.054989] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 536.054989] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 536.054989] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 536.054989] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] vm_ref = self.build_virtual_machine(instance, [ 536.054989] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 536.054989] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] vif_infos = vmwarevif.get_vif_info(self._session, [ 536.054989] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 536.055331] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] for vif in network_info: [ 536.055331] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 536.055331] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] return self._sync_wrapper(fn, *args, **kwargs) [ 536.055331] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 536.055331] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] self.wait() [ 536.055331] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 536.055331] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] self[:] = self._gt.wait() [ 536.055331] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 536.055331] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] return self._exit_event.wait() [ 536.055331] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 536.055331] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] result = hub.switch() [ 536.055331] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 536.055331] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] return self.greenlet.switch() [ 536.055657] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.055657] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] result = function(*args, **kwargs) [ 536.055657] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 536.055657] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] return func(*args, **kwargs) [ 536.055657] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 536.055657] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] raise e [ 536.055657] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 536.055657] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] nwinfo = self.network_api.allocate_for_instance( [ 536.055657] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 536.055657] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] created_port_ids = self._update_ports_for_instance( [ 536.055657] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 536.055657] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] with excutils.save_and_reraise_exception(): [ 536.055657] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.055966] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] self.force_reraise() [ 536.055966] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.055966] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] raise self.value [ 536.055966] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 536.055966] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] updated_port = self._update_port( [ 536.055966] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.055966] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] _ensure_no_port_binding_failure(port) [ 536.055966] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.055966] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] raise exception.PortBindingFailed(port_id=port['id']) [ 536.055966] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] nova.exception.PortBindingFailed: Binding failed for port 409365e0-c0a3-4367-add1-3803636b9262, please check neutron logs for more information. [ 536.055966] env[62569]: ERROR nova.compute.manager [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] [ 536.056252] env[62569]: DEBUG nova.compute.utils [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Binding failed for port 409365e0-c0a3-4367-add1-3803636b9262, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 536.057837] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.093s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.058708] env[62569]: INFO nova.compute.claims [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 536.063018] env[62569]: DEBUG nova.compute.manager [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Build of instance ff90249f-ae4b-4aeb-91ed-d8a41a942a5d was re-scheduled: Binding failed for port 409365e0-c0a3-4367-add1-3803636b9262, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 536.063018] env[62569]: DEBUG nova.compute.manager [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 536.063018] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Acquiring lock "refresh_cache-ff90249f-ae4b-4aeb-91ed-d8a41a942a5d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.063018] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Acquired lock "refresh_cache-ff90249f-ae4b-4aeb-91ed-d8a41a942a5d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.063540] env[62569]: DEBUG nova.network.neutron [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 536.079960] env[62569]: DEBUG nova.compute.manager [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 536.110060] env[62569]: DEBUG nova.virt.hardware [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 536.110299] env[62569]: DEBUG nova.virt.hardware [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 536.110565] env[62569]: DEBUG nova.virt.hardware [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 536.110932] env[62569]: DEBUG nova.virt.hardware [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 536.111483] env[62569]: DEBUG nova.virt.hardware [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 536.111483] env[62569]: DEBUG nova.virt.hardware [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 536.111617] env[62569]: DEBUG nova.virt.hardware [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 536.112041] env[62569]: DEBUG nova.virt.hardware [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 536.112282] env[62569]: DEBUG nova.virt.hardware [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 536.112495] env[62569]: DEBUG nova.virt.hardware [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 536.112863] env[62569]: DEBUG nova.virt.hardware [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 536.114512] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42f00ba9-8463-40a2-9708-51e518717a89 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.123894] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79dbe2c3-dd39-4d05-b7d9-ca1f69e157c1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.140282] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Instance VIF info [] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 536.148027] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Creating folder: Project (65d461cfdecd4bbfa22713a472043aaf). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 536.148027] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9fe69e4c-a9b8-4e9b-8c0e-27215244ef67 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.155998] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Created folder: Project (65d461cfdecd4bbfa22713a472043aaf) in parent group-v269330. [ 536.159686] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Creating folder: Instances. Parent ref: group-v269337. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 536.159686] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-17fcd4db-e854-4cf1-908e-eeb84d11bd35 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.165843] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Created folder: Instances in parent group-v269337. [ 536.166093] env[62569]: DEBUG oslo.service.loopingcall [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 536.166279] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 536.166474] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-84874185-915b-4b0e-bf6f-00556df1f9be {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.187030] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 536.187030] env[62569]: value = "task-1249808" [ 536.187030] env[62569]: _type = "Task" [ 536.187030] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 536.193521] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249808, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 536.503172] env[62569]: DEBUG nova.compute.manager [req-b2b3c1e7-cecd-4ac9-bfd0-128f07634ccb req-21ac3d74-9d2a-4c69-ab5f-a356b66a17c9 service nova] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Received event network-changed-ae242046-3aec-42af-a3cc-1839937c8497 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 536.503222] env[62569]: DEBUG nova.compute.manager [req-b2b3c1e7-cecd-4ac9-bfd0-128f07634ccb req-21ac3d74-9d2a-4c69-ab5f-a356b66a17c9 service nova] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Refreshing instance network info cache due to event network-changed-ae242046-3aec-42af-a3cc-1839937c8497. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 536.503417] env[62569]: DEBUG oslo_concurrency.lockutils [req-b2b3c1e7-cecd-4ac9-bfd0-128f07634ccb req-21ac3d74-9d2a-4c69-ab5f-a356b66a17c9 service nova] Acquiring lock "refresh_cache-e8b8b029-b20a-4af3-a5dc-1b032ce46f8d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.503849] env[62569]: DEBUG oslo_concurrency.lockutils [req-b2b3c1e7-cecd-4ac9-bfd0-128f07634ccb req-21ac3d74-9d2a-4c69-ab5f-a356b66a17c9 service nova] Acquired lock "refresh_cache-e8b8b029-b20a-4af3-a5dc-1b032ce46f8d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.504051] env[62569]: DEBUG nova.network.neutron [req-b2b3c1e7-cecd-4ac9-bfd0-128f07634ccb req-21ac3d74-9d2a-4c69-ab5f-a356b66a17c9 service nova] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Refreshing network info cache for port ae242046-3aec-42af-a3cc-1839937c8497 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 536.613553] env[62569]: DEBUG nova.network.neutron [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 536.617446] env[62569]: ERROR nova.compute.manager [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ae242046-3aec-42af-a3cc-1839937c8497, please check neutron logs for more information. [ 536.617446] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 536.617446] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 536.617446] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 536.617446] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 536.617446] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 536.617446] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 536.617446] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 536.617446] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.617446] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 536.617446] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.617446] env[62569]: ERROR nova.compute.manager raise self.value [ 536.617446] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 536.617446] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 536.617446] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.617446] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 536.618204] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.618204] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 536.618204] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ae242046-3aec-42af-a3cc-1839937c8497, please check neutron logs for more information. [ 536.618204] env[62569]: ERROR nova.compute.manager [ 536.618204] env[62569]: Traceback (most recent call last): [ 536.618204] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 536.618204] env[62569]: listener.cb(fileno) [ 536.618204] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.618204] env[62569]: result = function(*args, **kwargs) [ 536.618204] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 536.618204] env[62569]: return func(*args, **kwargs) [ 536.618204] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 536.618204] env[62569]: raise e [ 536.618204] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 536.618204] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 536.618204] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 536.618204] env[62569]: created_port_ids = self._update_ports_for_instance( [ 536.618204] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 536.618204] env[62569]: with excutils.save_and_reraise_exception(): [ 536.618204] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.618204] env[62569]: self.force_reraise() [ 536.618204] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.618204] env[62569]: raise self.value [ 536.618204] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 536.618204] env[62569]: updated_port = self._update_port( [ 536.618204] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.618204] env[62569]: _ensure_no_port_binding_failure(port) [ 536.618204] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.618204] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 536.620011] env[62569]: nova.exception.PortBindingFailed: Binding failed for port ae242046-3aec-42af-a3cc-1839937c8497, please check neutron logs for more information. [ 536.620011] env[62569]: Removing descriptor: 14 [ 536.620011] env[62569]: ERROR nova.compute.manager [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ae242046-3aec-42af-a3cc-1839937c8497, please check neutron logs for more information. [ 536.620011] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Traceback (most recent call last): [ 536.620011] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 536.620011] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] yield resources [ 536.620011] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 536.620011] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] self.driver.spawn(context, instance, image_meta, [ 536.620011] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 536.620011] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 536.620011] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 536.620011] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] vm_ref = self.build_virtual_machine(instance, [ 536.620643] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 536.620643] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] vif_infos = vmwarevif.get_vif_info(self._session, [ 536.620643] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 536.620643] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] for vif in network_info: [ 536.620643] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 536.620643] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] return self._sync_wrapper(fn, *args, **kwargs) [ 536.620643] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 536.620643] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] self.wait() [ 536.620643] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 536.620643] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] self[:] = self._gt.wait() [ 536.620643] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 536.620643] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] return self._exit_event.wait() [ 536.620643] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 536.621028] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] result = hub.switch() [ 536.621028] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 536.621028] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] return self.greenlet.switch() [ 536.621028] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.621028] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] result = function(*args, **kwargs) [ 536.621028] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 536.621028] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] return func(*args, **kwargs) [ 536.621028] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 536.621028] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] raise e [ 536.621028] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 536.621028] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] nwinfo = self.network_api.allocate_for_instance( [ 536.621028] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 536.621028] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] created_port_ids = self._update_ports_for_instance( [ 536.621394] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 536.621394] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] with excutils.save_and_reraise_exception(): [ 536.621394] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.621394] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] self.force_reraise() [ 536.621394] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.621394] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] raise self.value [ 536.621394] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 536.621394] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] updated_port = self._update_port( [ 536.621394] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.621394] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] _ensure_no_port_binding_failure(port) [ 536.621394] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.621394] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] raise exception.PortBindingFailed(port_id=port['id']) [ 536.621706] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] nova.exception.PortBindingFailed: Binding failed for port ae242046-3aec-42af-a3cc-1839937c8497, please check neutron logs for more information. [ 536.621706] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] [ 536.621706] env[62569]: INFO nova.compute.manager [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Terminating instance [ 536.699467] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249808, 'name': CreateVM_Task, 'duration_secs': 0.299238} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 536.699803] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 536.701110] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.701110] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.701110] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 536.701819] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64f5fb61-8560-46b6-b52e-61b75cc35547 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.707491] env[62569]: DEBUG oslo_vmware.api [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Waiting for the task: (returnval){ [ 536.707491] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52058cfb-59df-ada5-0701-25ea89dc9e60" [ 536.707491] env[62569]: _type = "Task" [ 536.707491] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 536.717602] env[62569]: DEBUG oslo_vmware.api [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52058cfb-59df-ada5-0701-25ea89dc9e60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 536.850649] env[62569]: DEBUG nova.network.neutron [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.059528] env[62569]: DEBUG nova.network.neutron [req-b2b3c1e7-cecd-4ac9-bfd0-128f07634ccb req-21ac3d74-9d2a-4c69-ab5f-a356b66a17c9 service nova] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 537.119862] env[62569]: DEBUG nova.compute.manager [req-cdb2e6f7-2e0c-4427-b0ae-1035a5fcb7f4 req-a227533f-8985-4cd8-83e3-e39699e8e56f service nova] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Received event network-changed-28017eb1-a7ac-4cfb-a089-ad527f6ed915 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 537.119952] env[62569]: DEBUG nova.compute.manager [req-cdb2e6f7-2e0c-4427-b0ae-1035a5fcb7f4 req-a227533f-8985-4cd8-83e3-e39699e8e56f service nova] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Refreshing instance network info cache due to event network-changed-28017eb1-a7ac-4cfb-a089-ad527f6ed915. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 537.121088] env[62569]: DEBUG oslo_concurrency.lockutils [req-cdb2e6f7-2e0c-4427-b0ae-1035a5fcb7f4 req-a227533f-8985-4cd8-83e3-e39699e8e56f service nova] Acquiring lock "refresh_cache-ba23eea7-422c-4e5d-b95e-f95589e41550" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.121236] env[62569]: DEBUG oslo_concurrency.lockutils [req-cdb2e6f7-2e0c-4427-b0ae-1035a5fcb7f4 req-a227533f-8985-4cd8-83e3-e39699e8e56f service nova] Acquired lock "refresh_cache-ba23eea7-422c-4e5d-b95e-f95589e41550" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.122140] env[62569]: DEBUG nova.network.neutron [req-cdb2e6f7-2e0c-4427-b0ae-1035a5fcb7f4 req-a227533f-8985-4cd8-83e3-e39699e8e56f service nova] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Refreshing network info cache for port 28017eb1-a7ac-4cfb-a089-ad527f6ed915 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 537.123281] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Acquiring lock "refresh_cache-e8b8b029-b20a-4af3-a5dc-1b032ce46f8d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.222418] env[62569]: DEBUG oslo_vmware.api [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52058cfb-59df-ada5-0701-25ea89dc9e60, 'name': SearchDatastore_Task, 'duration_secs': 0.009499} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 537.222418] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.222418] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 537.222617] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.222738] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.223090] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 537.223374] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b87590cc-ede2-44cc-b6c0-8f104aba7a1f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.225614] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Acquiring lock "17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.225828] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Lock "17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.227270] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Acquiring lock "17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.227270] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Lock "17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.227270] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Lock "17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.236508] env[62569]: INFO nova.compute.manager [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Terminating instance [ 537.247067] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 537.247304] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 537.248923] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2a02c37-d4df-4e53-bade-281d96bf1cde {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.263149] env[62569]: DEBUG oslo_vmware.api [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Waiting for the task: (returnval){ [ 537.263149] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5217f18c-4694-eebf-4abe-6a3b1d3b26e1" [ 537.263149] env[62569]: _type = "Task" [ 537.263149] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.277707] env[62569]: DEBUG oslo_vmware.api [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5217f18c-4694-eebf-4abe-6a3b1d3b26e1, 'name': SearchDatastore_Task, 'duration_secs': 0.014305} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 537.281219] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-839b821e-cf6f-474a-8ae8-f55ac7409fd0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.287695] env[62569]: DEBUG oslo_vmware.api [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Waiting for the task: (returnval){ [ 537.287695] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52e2e99f-0b3e-480e-744f-c081e5fa0624" [ 537.287695] env[62569]: _type = "Task" [ 537.287695] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.297030] env[62569]: DEBUG oslo_vmware.api [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52e2e99f-0b3e-480e-744f-c081e5fa0624, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.329232] env[62569]: DEBUG nova.network.neutron [req-b2b3c1e7-cecd-4ac9-bfd0-128f07634ccb req-21ac3d74-9d2a-4c69-ab5f-a356b66a17c9 service nova] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.353525] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Releasing lock "refresh_cache-ff90249f-ae4b-4aeb-91ed-d8a41a942a5d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.356040] env[62569]: DEBUG nova.compute.manager [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 537.356040] env[62569]: DEBUG nova.compute.manager [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 537.356040] env[62569]: DEBUG nova.network.neutron [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 537.394830] env[62569]: DEBUG nova.network.neutron [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 537.488307] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3848e7f5-320f-487b-9126-03f081532041 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.500677] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48871bf6-c550-4ddc-a1a5-3ca9426cc79a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.545811] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d16cc8a-7ed0-4567-8865-48e03d6148f7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.554538] env[62569]: ERROR nova.compute.manager [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 28017eb1-a7ac-4cfb-a089-ad527f6ed915, please check neutron logs for more information. [ 537.554538] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 537.554538] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 537.554538] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 537.554538] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 537.554538] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 537.554538] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 537.554538] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 537.554538] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.554538] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 537.554538] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.554538] env[62569]: ERROR nova.compute.manager raise self.value [ 537.554538] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 537.554538] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 537.554538] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.554538] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 537.555098] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.555098] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 537.555098] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 28017eb1-a7ac-4cfb-a089-ad527f6ed915, please check neutron logs for more information. [ 537.555098] env[62569]: ERROR nova.compute.manager [ 537.555098] env[62569]: Traceback (most recent call last): [ 537.555098] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 537.555098] env[62569]: listener.cb(fileno) [ 537.555098] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.555098] env[62569]: result = function(*args, **kwargs) [ 537.555098] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 537.555098] env[62569]: return func(*args, **kwargs) [ 537.555098] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 537.555098] env[62569]: raise e [ 537.555098] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 537.555098] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 537.555098] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 537.555098] env[62569]: created_port_ids = self._update_ports_for_instance( [ 537.555098] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 537.555098] env[62569]: with excutils.save_and_reraise_exception(): [ 537.555098] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.555098] env[62569]: self.force_reraise() [ 537.555098] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.555098] env[62569]: raise self.value [ 537.555098] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 537.555098] env[62569]: updated_port = self._update_port( [ 537.555098] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.555098] env[62569]: _ensure_no_port_binding_failure(port) [ 537.555098] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.555098] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 537.555825] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 28017eb1-a7ac-4cfb-a089-ad527f6ed915, please check neutron logs for more information. [ 537.555825] env[62569]: Removing descriptor: 16 [ 537.555825] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc237a8-8d88-40d4-87b3-99b77333ccec {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.564604] env[62569]: ERROR nova.compute.manager [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 28017eb1-a7ac-4cfb-a089-ad527f6ed915, please check neutron logs for more information. [ 537.564604] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Traceback (most recent call last): [ 537.564604] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 537.564604] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] yield resources [ 537.564604] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 537.564604] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] self.driver.spawn(context, instance, image_meta, [ 537.564604] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 537.564604] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] self._vmops.spawn(context, instance, image_meta, injected_files, [ 537.564604] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 537.564604] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] vm_ref = self.build_virtual_machine(instance, [ 537.564604] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 537.565239] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] vif_infos = vmwarevif.get_vif_info(self._session, [ 537.565239] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 537.565239] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] for vif in network_info: [ 537.565239] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 537.565239] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] return self._sync_wrapper(fn, *args, **kwargs) [ 537.565239] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 537.565239] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] self.wait() [ 537.565239] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 537.565239] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] self[:] = self._gt.wait() [ 537.565239] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 537.565239] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] return self._exit_event.wait() [ 537.565239] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 537.565239] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] result = hub.switch() [ 537.565820] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 537.565820] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] return self.greenlet.switch() [ 537.565820] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.565820] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] result = function(*args, **kwargs) [ 537.565820] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 537.565820] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] return func(*args, **kwargs) [ 537.565820] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 537.565820] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] raise e [ 537.565820] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 537.565820] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] nwinfo = self.network_api.allocate_for_instance( [ 537.565820] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 537.565820] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] created_port_ids = self._update_ports_for_instance( [ 537.565820] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 537.567485] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] with excutils.save_and_reraise_exception(): [ 537.567485] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.567485] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] self.force_reraise() [ 537.567485] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.567485] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] raise self.value [ 537.567485] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 537.567485] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] updated_port = self._update_port( [ 537.567485] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.567485] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] _ensure_no_port_binding_failure(port) [ 537.567485] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.567485] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] raise exception.PortBindingFailed(port_id=port['id']) [ 537.567485] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] nova.exception.PortBindingFailed: Binding failed for port 28017eb1-a7ac-4cfb-a089-ad527f6ed915, please check neutron logs for more information. [ 537.567485] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] [ 537.572693] env[62569]: INFO nova.compute.manager [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Terminating instance [ 537.579235] env[62569]: DEBUG nova.compute.provider_tree [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 537.677786] env[62569]: DEBUG nova.network.neutron [req-cdb2e6f7-2e0c-4427-b0ae-1035a5fcb7f4 req-a227533f-8985-4cd8-83e3-e39699e8e56f service nova] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 537.740994] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Acquiring lock "refresh_cache-17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.741431] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Acquired lock "refresh_cache-17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.741575] env[62569]: DEBUG nova.network.neutron [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 537.804564] env[62569]: DEBUG oslo_vmware.api [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52e2e99f-0b3e-480e-744f-c081e5fa0624, 'name': SearchDatastore_Task, 'duration_secs': 0.011241} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 537.809130] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.809130] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 6e35fe1c-05bc-4dcc-b7ec-e9998a647739/6e35fe1c-05bc-4dcc-b7ec-e9998a647739.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 537.809130] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0032a067-6370-4fe9-aebd-76f7f61cdece {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.815942] env[62569]: DEBUG oslo_vmware.api [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Waiting for the task: (returnval){ [ 537.815942] env[62569]: value = "task-1249810" [ 537.815942] env[62569]: _type = "Task" [ 537.815942] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.827062] env[62569]: DEBUG oslo_vmware.api [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249810, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.833132] env[62569]: DEBUG oslo_concurrency.lockutils [req-b2b3c1e7-cecd-4ac9-bfd0-128f07634ccb req-21ac3d74-9d2a-4c69-ab5f-a356b66a17c9 service nova] Releasing lock "refresh_cache-e8b8b029-b20a-4af3-a5dc-1b032ce46f8d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.834599] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Acquired lock "refresh_cache-e8b8b029-b20a-4af3-a5dc-1b032ce46f8d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.834599] env[62569]: DEBUG nova.network.neutron [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 537.897945] env[62569]: DEBUG nova.network.neutron [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.929850] env[62569]: DEBUG nova.network.neutron [req-cdb2e6f7-2e0c-4427-b0ae-1035a5fcb7f4 req-a227533f-8985-4cd8-83e3-e39699e8e56f service nova] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.082515] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Acquiring lock "refresh_cache-ba23eea7-422c-4e5d-b95e-f95589e41550" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.084725] env[62569]: DEBUG nova.scheduler.client.report [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 538.305739] env[62569]: DEBUG nova.network.neutron [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.330151] env[62569]: DEBUG oslo_vmware.api [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249810, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465047} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.330446] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 6e35fe1c-05bc-4dcc-b7ec-e9998a647739/6e35fe1c-05bc-4dcc-b7ec-e9998a647739.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 538.331173] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 538.331173] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7f94f3e2-a4d6-485c-862c-c89820eb3404 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.339946] env[62569]: DEBUG oslo_vmware.api [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Waiting for the task: (returnval){ [ 538.339946] env[62569]: value = "task-1249811" [ 538.339946] env[62569]: _type = "Task" [ 538.339946] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.352592] env[62569]: DEBUG oslo_vmware.api [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249811, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.377257] env[62569]: DEBUG nova.network.neutron [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.404472] env[62569]: INFO nova.compute.manager [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] Took 1.05 seconds to deallocate network for instance. [ 538.438883] env[62569]: DEBUG oslo_concurrency.lockutils [req-cdb2e6f7-2e0c-4427-b0ae-1035a5fcb7f4 req-a227533f-8985-4cd8-83e3-e39699e8e56f service nova] Releasing lock "refresh_cache-ba23eea7-422c-4e5d-b95e-f95589e41550" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.439323] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Acquired lock "refresh_cache-ba23eea7-422c-4e5d-b95e-f95589e41550" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.440760] env[62569]: DEBUG nova.network.neutron [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 538.497341] env[62569]: DEBUG nova.network.neutron [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.515243] env[62569]: DEBUG nova.network.neutron [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.591969] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.535s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 538.592557] env[62569]: DEBUG nova.compute.manager [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 538.599936] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 13.581s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.600158] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 538.600346] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62569) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 538.604018] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.901s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.606720] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1d391d4-58ac-4c86-8cd4-8c7c86dee9c6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.622410] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8615c5a-7a6d-472e-8d15-922db49645d4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.638774] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8098c4b1-3926-4954-a1a9-a04bcd4e6a63 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.658131] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9160c4ca-7e7a-4c1e-b529-66836f5b5293 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.696398] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181543MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=62569) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 538.696398] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.856186] env[62569]: DEBUG oslo_vmware.api [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249811, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061398} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.856501] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 538.857348] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f91500ca-0063-407c-b02a-4b66339d1c0a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.880830] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Reconfiguring VM instance instance-0000000c to attach disk [datastore1] 6e35fe1c-05bc-4dcc-b7ec-e9998a647739/6e35fe1c-05bc-4dcc-b7ec-e9998a647739.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 538.881424] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-842cd1e8-e42f-484d-9ba6-c5088d721ef9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.906209] env[62569]: DEBUG oslo_vmware.api [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Waiting for the task: (returnval){ [ 538.906209] env[62569]: value = "task-1249812" [ 538.906209] env[62569]: _type = "Task" [ 538.906209] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.925376] env[62569]: DEBUG oslo_vmware.api [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249812, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.000533] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Releasing lock "refresh_cache-e8b8b029-b20a-4af3-a5dc-1b032ce46f8d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.001183] env[62569]: DEBUG nova.compute.manager [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 539.001294] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 539.001700] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5fc06085-c443-4bb3-a73b-c6f7c9cffc51 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.018748] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-027fa7c7-7e0a-44c5-97ac-5e3a228b9e59 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.033306] env[62569]: DEBUG nova.network.neutron [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.035239] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Releasing lock "refresh_cache-17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.035651] env[62569]: DEBUG nova.compute.manager [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 539.035988] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 539.037266] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e7ce808-e1df-47fa-9501-5e0576c5ff93 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.053023] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 539.053023] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ea1617ec-16ec-4b6a-9ed0-fc60d4f92457 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.060986] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e8b8b029-b20a-4af3-a5dc-1b032ce46f8d could not be found. [ 539.061456] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 539.061677] env[62569]: INFO nova.compute.manager [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Took 0.06 seconds to destroy the instance on the hypervisor. [ 539.061942] env[62569]: DEBUG oslo.service.loopingcall [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 539.063507] env[62569]: DEBUG nova.compute.manager [-] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 539.063650] env[62569]: DEBUG nova.network.neutron [-] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 539.069485] env[62569]: DEBUG oslo_vmware.api [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Waiting for the task: (returnval){ [ 539.069485] env[62569]: value = "task-1249814" [ 539.069485] env[62569]: _type = "Task" [ 539.069485] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.080210] env[62569]: DEBUG oslo_vmware.api [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249814, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.104892] env[62569]: DEBUG nova.network.neutron [-] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.107354] env[62569]: DEBUG nova.compute.utils [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 539.108789] env[62569]: DEBUG nova.compute.manager [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 539.108944] env[62569]: DEBUG nova.network.neutron [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 539.229394] env[62569]: DEBUG nova.policy [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4b971531d3ff428cbea1742c1629d8ca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c917db0ac9124b9aa2be1e0a0bffb662', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 539.381743] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Acquiring lock "2a61b4cd-a15c-4064-9351-326dd777603b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.382000] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Lock "2a61b4cd-a15c-4064-9351-326dd777603b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.418130] env[62569]: DEBUG oslo_vmware.api [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249812, 'name': ReconfigVM_Task, 'duration_secs': 0.266127} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.418765] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Reconfigured VM instance instance-0000000c to attach disk [datastore1] 6e35fe1c-05bc-4dcc-b7ec-e9998a647739/6e35fe1c-05bc-4dcc-b7ec-e9998a647739.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 539.419095] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-abea27d2-0b62-49bd-acd7-92084041e972 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.430252] env[62569]: DEBUG oslo_vmware.api [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Waiting for the task: (returnval){ [ 539.430252] env[62569]: value = "task-1249815" [ 539.430252] env[62569]: _type = "Task" [ 539.430252] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.438181] env[62569]: DEBUG oslo_vmware.api [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249815, 'name': Rename_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.458775] env[62569]: INFO nova.scheduler.client.report [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Deleted allocations for instance ff90249f-ae4b-4aeb-91ed-d8a41a942a5d [ 539.528635] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb6ffd5a-c797-4f02-854f-f95f34355706 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.537998] env[62569]: DEBUG nova.network.neutron [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.539869] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f0e78e7-0ec3-4967-a25a-ac95c1bcce96 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.585229] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-921d9c80-66d7-4a54-bf60-e63263adf968 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.592690] env[62569]: DEBUG nova.compute.manager [req-b02ad4cb-44a7-4b1b-9f33-1c8ea681b6a9 req-8f23dd9f-cc86-4df7-a7d8-fcbbac2ce6c2 service nova] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Received event network-vif-deleted-ae242046-3aec-42af-a3cc-1839937c8497 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 539.598392] env[62569]: DEBUG oslo_vmware.api [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249814, 'name': PowerOffVM_Task, 'duration_secs': 0.147726} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.601175] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 539.601175] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 539.601175] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3a484f56-2e37-4ba6-b0c9-848e84da1c83 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.603906] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8828cf63-f753-44f6-bb81-a653a1b184e6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.609207] env[62569]: DEBUG nova.network.neutron [-] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.611927] env[62569]: DEBUG nova.compute.manager [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 539.624181] env[62569]: DEBUG nova.compute.provider_tree [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 539.633774] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 539.633956] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 539.634225] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Deleting the datastore file [datastore1] 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 539.634869] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3e8aa728-d45e-47cb-a7a1-e6fb19c6879f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.644856] env[62569]: DEBUG oslo_vmware.api [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Waiting for the task: (returnval){ [ 539.644856] env[62569]: value = "task-1249817" [ 539.644856] env[62569]: _type = "Task" [ 539.644856] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.653618] env[62569]: DEBUG oslo_vmware.api [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249817, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.823161] env[62569]: DEBUG nova.network.neutron [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Successfully created port: 2a118c7a-8fdc-4948-9ec7-facf3bf84bd8 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 539.946511] env[62569]: DEBUG oslo_vmware.api [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249815, 'name': Rename_Task, 'duration_secs': 0.142801} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.946511] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 539.946511] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c6de6031-11f6-4645-ad96-0cad07430d47 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.953599] env[62569]: DEBUG oslo_vmware.api [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Waiting for the task: (returnval){ [ 539.953599] env[62569]: value = "task-1249818" [ 539.953599] env[62569]: _type = "Task" [ 539.953599] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.963694] env[62569]: DEBUG oslo_vmware.api [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249818, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.975032] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8b4db70c-ccde-44bf-af5c-addc64fb4744 tempest-ServerExternalEventsTest-1415515260 tempest-ServerExternalEventsTest-1415515260-project-member] Lock "ff90249f-ae4b-4aeb-91ed-d8a41a942a5d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.373s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.976578] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "ff90249f-ae4b-4aeb-91ed-d8a41a942a5d" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 15.464s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.976818] env[62569]: INFO nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: ff90249f-ae4b-4aeb-91ed-d8a41a942a5d] During sync_power_state the instance has a pending task (spawning). Skip. [ 539.977016] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "ff90249f-ae4b-4aeb-91ed-d8a41a942a5d" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.045857] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Releasing lock "refresh_cache-ba23eea7-422c-4e5d-b95e-f95589e41550" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.046844] env[62569]: DEBUG nova.compute.manager [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 540.046844] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 540.046964] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7f8f6cb0-db91-41bd-a12c-2c259086eec5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.055958] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc971993-0bb8-45f9-abdf-c12b7a53a1c4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.086879] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ba23eea7-422c-4e5d-b95e-f95589e41550 could not be found. [ 540.087182] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 540.087345] env[62569]: INFO nova.compute.manager [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Took 0.04 seconds to destroy the instance on the hypervisor. [ 540.087754] env[62569]: DEBUG oslo.service.loopingcall [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 540.088010] env[62569]: DEBUG nova.compute.manager [-] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 540.088115] env[62569]: DEBUG nova.network.neutron [-] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 540.115571] env[62569]: INFO nova.compute.manager [-] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Took 1.05 seconds to deallocate network for instance. [ 540.118135] env[62569]: DEBUG nova.compute.claims [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 540.118788] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.128436] env[62569]: DEBUG nova.scheduler.client.report [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 540.161816] env[62569]: DEBUG oslo_vmware.api [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Task: {'id': task-1249817, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.091452} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.162087] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 540.162277] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 540.162452] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 540.162609] env[62569]: INFO nova.compute.manager [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Took 1.13 seconds to destroy the instance on the hypervisor. [ 540.162834] env[62569]: DEBUG oslo.service.loopingcall [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 540.163139] env[62569]: DEBUG nova.compute.manager [-] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 540.163228] env[62569]: DEBUG nova.network.neutron [-] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 540.168224] env[62569]: DEBUG nova.network.neutron [-] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 540.212115] env[62569]: DEBUG nova.network.neutron [-] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 540.463808] env[62569]: DEBUG oslo_vmware.api [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249818, 'name': PowerOnVM_Task, 'duration_secs': 0.442719} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.464440] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 540.464693] env[62569]: INFO nova.compute.manager [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Took 4.38 seconds to spawn the instance on the hypervisor. [ 540.464876] env[62569]: DEBUG nova.compute.manager [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 540.465731] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f6d359b-c03b-478b-b53f-0f42a593e78b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.478351] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 540.636595] env[62569]: DEBUG nova.compute.manager [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 540.641462] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.041s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.648950] env[62569]: ERROR nova.compute.manager [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 95c9ce89-4e44-45c7-8dcb-4c285a137ff2, please check neutron logs for more information. [ 540.648950] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Traceback (most recent call last): [ 540.648950] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 540.648950] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] self.driver.spawn(context, instance, image_meta, [ 540.648950] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 540.648950] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] self._vmops.spawn(context, instance, image_meta, injected_files, [ 540.648950] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 540.648950] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] vm_ref = self.build_virtual_machine(instance, [ 540.648950] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 540.648950] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] vif_infos = vmwarevif.get_vif_info(self._session, [ 540.648950] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 540.649782] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] for vif in network_info: [ 540.649782] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 540.649782] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] return self._sync_wrapper(fn, *args, **kwargs) [ 540.649782] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 540.649782] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] self.wait() [ 540.649782] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 540.649782] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] self[:] = self._gt.wait() [ 540.649782] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 540.649782] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] return self._exit_event.wait() [ 540.649782] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 540.649782] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] result = hub.switch() [ 540.649782] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 540.649782] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] return self.greenlet.switch() [ 540.650205] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.650205] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] result = function(*args, **kwargs) [ 540.650205] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 540.650205] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] return func(*args, **kwargs) [ 540.650205] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 540.650205] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] raise e [ 540.650205] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 540.650205] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] nwinfo = self.network_api.allocate_for_instance( [ 540.650205] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 540.650205] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] created_port_ids = self._update_ports_for_instance( [ 540.650205] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 540.650205] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] with excutils.save_and_reraise_exception(): [ 540.650205] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.650658] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] self.force_reraise() [ 540.650658] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.650658] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] raise self.value [ 540.650658] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 540.650658] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] updated_port = self._update_port( [ 540.650658] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.650658] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] _ensure_no_port_binding_failure(port) [ 540.650658] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.650658] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] raise exception.PortBindingFailed(port_id=port['id']) [ 540.650658] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] nova.exception.PortBindingFailed: Binding failed for port 95c9ce89-4e44-45c7-8dcb-4c285a137ff2, please check neutron logs for more information. [ 540.650658] env[62569]: ERROR nova.compute.manager [instance: 832ec210-bb14-4a52-9634-50b7709a9976] [ 540.650952] env[62569]: DEBUG nova.compute.utils [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Binding failed for port 95c9ce89-4e44-45c7-8dcb-4c285a137ff2, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 540.653645] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.199s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.653645] env[62569]: INFO nova.compute.claims [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 540.662654] env[62569]: DEBUG nova.compute.manager [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Build of instance 832ec210-bb14-4a52-9634-50b7709a9976 was re-scheduled: Binding failed for port 95c9ce89-4e44-45c7-8dcb-4c285a137ff2, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 540.662654] env[62569]: DEBUG nova.compute.manager [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 540.662944] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Acquiring lock "refresh_cache-832ec210-bb14-4a52-9634-50b7709a9976" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.666019] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Acquired lock "refresh_cache-832ec210-bb14-4a52-9634-50b7709a9976" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.666019] env[62569]: DEBUG nova.network.neutron [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 540.670824] env[62569]: DEBUG nova.network.neutron [-] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.685755] env[62569]: DEBUG nova.virt.hardware [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 540.685755] env[62569]: DEBUG nova.virt.hardware [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 540.685755] env[62569]: DEBUG nova.virt.hardware [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 540.686021] env[62569]: DEBUG nova.virt.hardware [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 540.686021] env[62569]: DEBUG nova.virt.hardware [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 540.686021] env[62569]: DEBUG nova.virt.hardware [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 540.686021] env[62569]: DEBUG nova.virt.hardware [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 540.686021] env[62569]: DEBUG nova.virt.hardware [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 540.690183] env[62569]: DEBUG nova.virt.hardware [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 540.690183] env[62569]: DEBUG nova.virt.hardware [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 540.690183] env[62569]: DEBUG nova.virt.hardware [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 540.690665] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cbc65e4-b641-42aa-8be7-c255e3bed7d0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.701737] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51691a37-302a-4183-aab5-1f7f87c633da {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.717445] env[62569]: DEBUG nova.network.neutron [-] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.004410] env[62569]: INFO nova.compute.manager [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Took 18.24 seconds to build instance. [ 541.010955] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.177766] env[62569]: INFO nova.compute.manager [-] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Took 1.09 seconds to deallocate network for instance. [ 541.180123] env[62569]: DEBUG nova.compute.claims [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 541.180363] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.207653] env[62569]: DEBUG nova.network.neutron [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 541.221410] env[62569]: INFO nova.compute.manager [-] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Took 1.06 seconds to deallocate network for instance. [ 541.394637] env[62569]: DEBUG nova.compute.manager [req-de7bfab1-9d85-4fb5-b1a7-3129c2965bbf req-4bc0e961-7cdf-4401-a5d2-242628e33b3f service nova] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Received event network-vif-deleted-28017eb1-a7ac-4cfb-a089-ad527f6ed915 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 541.507382] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d5a1da4d-37a7-404f-967c-60ba94a49b75 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Lock "6e35fe1c-05bc-4dcc-b7ec-e9998a647739" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.748s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.576610] env[62569]: DEBUG nova.network.neutron [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.728120] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.948638] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Acquiring lock "76bddf93-ca4f-4987-a446-b369801ebd93" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.948638] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Lock "76bddf93-ca4f-4987-a446-b369801ebd93" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.010805] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 542.014330] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6e4ea2d-08d9-450e-9d04-42f0914f8584 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.022694] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36c2c2c-7315-4235-bb25-1409a410a105 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.055807] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aa1f061-dde8-4f2f-9dc5-e5597c2c26ce {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.064310] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5986d3ab-9732-44ea-b889-30e9f02699c5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.078742] env[62569]: DEBUG nova.compute.provider_tree [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 542.080714] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Releasing lock "refresh_cache-832ec210-bb14-4a52-9634-50b7709a9976" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.080883] env[62569]: DEBUG nova.compute.manager [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 542.081280] env[62569]: DEBUG nova.compute.manager [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 542.082694] env[62569]: DEBUG nova.network.neutron [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 542.179803] env[62569]: DEBUG nova.network.neutron [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 542.537985] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.585898] env[62569]: DEBUG nova.scheduler.client.report [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 542.684268] env[62569]: DEBUG nova.network.neutron [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.011491] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Acquiring lock "4b9e4f73-4682-4170-b71e-1d3d1d860577" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.011491] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Lock "4b9e4f73-4682-4170-b71e-1d3d1d860577" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.093875] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.441s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.094273] env[62569]: DEBUG nova.compute.manager [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 543.098745] env[62569]: DEBUG oslo_concurrency.lockutils [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.052s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.188149] env[62569]: INFO nova.compute.manager [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] Took 1.11 seconds to deallocate network for instance. [ 543.467455] env[62569]: INFO nova.compute.manager [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Rebuilding instance [ 543.544095] env[62569]: DEBUG nova.compute.manager [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 543.545072] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cbcc66a-921e-4c74-a119-897d614f75a2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.609347] env[62569]: DEBUG nova.compute.utils [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 543.610759] env[62569]: DEBUG nova.compute.manager [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 543.610919] env[62569]: DEBUG nova.network.neutron [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 543.832875] env[62569]: DEBUG nova.policy [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e28265bc24244d5fa7e1c9d1f6ae2d26', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '454bd8bed8ed476abe622b2715653f72', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 543.834293] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "898f69fe-737a-44cb-9c43-965f1f8685f7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.834524] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "898f69fe-737a-44cb-9c43-965f1f8685f7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.050174] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5cf0fee-e3ab-4ba3-bf8c-6977a4c94ffc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.058702] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b008fb1-ec33-404e-8d17-52b95ffc6d9e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.097746] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-223c7f42-8990-44ab-8cc1-d1ba50e8a6ff {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.106586] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5d6794c-eadc-4ee8-b2aa-657b3f3e9726 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.127380] env[62569]: DEBUG nova.compute.manager [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 544.130602] env[62569]: DEBUG nova.compute.provider_tree [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 544.246630] env[62569]: INFO nova.scheduler.client.report [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Deleted allocations for instance 832ec210-bb14-4a52-9634-50b7709a9976 [ 544.568569] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 544.568828] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dc8bcdb5-edca-4021-862a-69179055ab1e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.579237] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Waiting for the task: (returnval){ [ 544.579237] env[62569]: value = "task-1249821" [ 544.579237] env[62569]: _type = "Task" [ 544.579237] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.591581] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249821, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.635881] env[62569]: DEBUG nova.scheduler.client.report [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 544.760845] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff396976-de68-494f-bf56-1c709af44de3 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Lock "832ec210-bb14-4a52-9634-50b7709a9976" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.700s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.760845] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "832ec210-bb14-4a52-9634-50b7709a9976" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 20.247s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.761341] env[62569]: INFO nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 832ec210-bb14-4a52-9634-50b7709a9976] During sync_power_state the instance has a pending task (spawning). Skip. [ 544.761798] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "832ec210-bb14-4a52-9634-50b7709a9976" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.002s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.912992] env[62569]: DEBUG nova.compute.manager [req-9c7f350a-5520-4d61-ba26-2bba5c4fc426 req-fcf72db7-1ea7-4699-867b-879868d0ad2c service nova] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Received event network-changed-2a118c7a-8fdc-4948-9ec7-facf3bf84bd8 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 544.913208] env[62569]: DEBUG nova.compute.manager [req-9c7f350a-5520-4d61-ba26-2bba5c4fc426 req-fcf72db7-1ea7-4699-867b-879868d0ad2c service nova] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Refreshing instance network info cache due to event network-changed-2a118c7a-8fdc-4948-9ec7-facf3bf84bd8. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 544.913431] env[62569]: DEBUG oslo_concurrency.lockutils [req-9c7f350a-5520-4d61-ba26-2bba5c4fc426 req-fcf72db7-1ea7-4699-867b-879868d0ad2c service nova] Acquiring lock "refresh_cache-9fa91603-f812-4d93-bef1-58e455f2c2e8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.913540] env[62569]: DEBUG oslo_concurrency.lockutils [req-9c7f350a-5520-4d61-ba26-2bba5c4fc426 req-fcf72db7-1ea7-4699-867b-879868d0ad2c service nova] Acquired lock "refresh_cache-9fa91603-f812-4d93-bef1-58e455f2c2e8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.913690] env[62569]: DEBUG nova.network.neutron [req-9c7f350a-5520-4d61-ba26-2bba5c4fc426 req-fcf72db7-1ea7-4699-867b-879868d0ad2c service nova] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Refreshing network info cache for port 2a118c7a-8fdc-4948-9ec7-facf3bf84bd8 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 545.092987] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249821, 'name': PowerOffVM_Task, 'duration_secs': 0.200511} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.093287] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 545.095036] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 545.098230] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-636870fb-fe33-4ccf-8641-b075bc2b4cbd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.108064] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 545.108390] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-416c0225-d8c6-4b61-aa74-540938f92399 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.136928] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 545.136928] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 545.136928] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Deleting the datastore file [datastore1] 6e35fe1c-05bc-4dcc-b7ec-e9998a647739 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 545.136928] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4e8a7313-9b40-4d8c-a990-ae4801e6d226 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.140923] env[62569]: DEBUG nova.compute.manager [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 545.145972] env[62569]: DEBUG oslo_concurrency.lockutils [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.048s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.146700] env[62569]: ERROR nova.compute.manager [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2077b280-333c-4da3-90fa-8a1a54c28e44, please check neutron logs for more information. [ 545.146700] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Traceback (most recent call last): [ 545.146700] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 545.146700] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] self.driver.spawn(context, instance, image_meta, [ 545.146700] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 545.146700] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] self._vmops.spawn(context, instance, image_meta, injected_files, [ 545.146700] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 545.146700] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] vm_ref = self.build_virtual_machine(instance, [ 545.146700] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 545.146700] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] vif_infos = vmwarevif.get_vif_info(self._session, [ 545.146700] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 545.147846] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] for vif in network_info: [ 545.147846] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 545.147846] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] return self._sync_wrapper(fn, *args, **kwargs) [ 545.147846] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 545.147846] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] self.wait() [ 545.147846] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 545.147846] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] self[:] = self._gt.wait() [ 545.147846] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 545.147846] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] return self._exit_event.wait() [ 545.147846] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 545.147846] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] result = hub.switch() [ 545.147846] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 545.147846] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] return self.greenlet.switch() [ 545.148235] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.148235] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] result = function(*args, **kwargs) [ 545.148235] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 545.148235] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] return func(*args, **kwargs) [ 545.148235] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 545.148235] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] raise e [ 545.148235] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 545.148235] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] nwinfo = self.network_api.allocate_for_instance( [ 545.148235] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 545.148235] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] created_port_ids = self._update_ports_for_instance( [ 545.148235] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 545.148235] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] with excutils.save_and_reraise_exception(): [ 545.148235] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.148583] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] self.force_reraise() [ 545.148583] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.148583] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] raise self.value [ 545.148583] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 545.148583] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] updated_port = self._update_port( [ 545.148583] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.148583] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] _ensure_no_port_binding_failure(port) [ 545.148583] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.148583] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] raise exception.PortBindingFailed(port_id=port['id']) [ 545.148583] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] nova.exception.PortBindingFailed: Binding failed for port 2077b280-333c-4da3-90fa-8a1a54c28e44, please check neutron logs for more information. [ 545.148583] env[62569]: ERROR nova.compute.manager [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] [ 545.148902] env[62569]: DEBUG nova.compute.utils [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Binding failed for port 2077b280-333c-4da3-90fa-8a1a54c28e44, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 545.149017] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.193s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.162534] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Waiting for the task: (returnval){ [ 545.162534] env[62569]: value = "task-1249823" [ 545.162534] env[62569]: _type = "Task" [ 545.162534] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.169772] env[62569]: DEBUG nova.compute.manager [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Build of instance 597c9673-bbd4-4dd0-829a-c05f8667c951 was re-scheduled: Binding failed for port 2077b280-333c-4da3-90fa-8a1a54c28e44, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 545.171031] env[62569]: DEBUG nova.compute.manager [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 545.171031] env[62569]: DEBUG oslo_concurrency.lockutils [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Acquiring lock "refresh_cache-597c9673-bbd4-4dd0-829a-c05f8667c951" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.171031] env[62569]: DEBUG oslo_concurrency.lockutils [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Acquired lock "refresh_cache-597c9673-bbd4-4dd0-829a-c05f8667c951" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.171031] env[62569]: DEBUG nova.network.neutron [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 545.180738] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249823, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.193117] env[62569]: DEBUG nova.virt.hardware [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 545.196679] env[62569]: DEBUG nova.virt.hardware [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 545.196679] env[62569]: DEBUG nova.virt.hardware [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 545.196679] env[62569]: DEBUG nova.virt.hardware [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 545.196679] env[62569]: DEBUG nova.virt.hardware [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 545.196679] env[62569]: DEBUG nova.virt.hardware [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 545.197081] env[62569]: DEBUG nova.virt.hardware [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 545.197081] env[62569]: DEBUG nova.virt.hardware [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 545.197081] env[62569]: DEBUG nova.virt.hardware [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 545.197081] env[62569]: DEBUG nova.virt.hardware [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 545.197081] env[62569]: DEBUG nova.virt.hardware [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 545.197282] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dfdf91e-ae8a-4887-bb68-b26da4021500 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.206537] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe534cb2-c376-481e-9841-63f5039c34f2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.227098] env[62569]: ERROR nova.compute.manager [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2a118c7a-8fdc-4948-9ec7-facf3bf84bd8, please check neutron logs for more information. [ 545.227098] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 545.227098] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 545.227098] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 545.227098] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 545.227098] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 545.227098] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 545.227098] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 545.227098] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.227098] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 545.227098] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.227098] env[62569]: ERROR nova.compute.manager raise self.value [ 545.227098] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 545.227098] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 545.227098] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.227098] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 545.227614] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.227614] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 545.227614] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2a118c7a-8fdc-4948-9ec7-facf3bf84bd8, please check neutron logs for more information. [ 545.227614] env[62569]: ERROR nova.compute.manager [ 545.227614] env[62569]: Traceback (most recent call last): [ 545.227614] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 545.227614] env[62569]: listener.cb(fileno) [ 545.227614] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.227614] env[62569]: result = function(*args, **kwargs) [ 545.227614] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 545.227614] env[62569]: return func(*args, **kwargs) [ 545.227614] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 545.227614] env[62569]: raise e [ 545.227614] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 545.227614] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 545.227614] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 545.227614] env[62569]: created_port_ids = self._update_ports_for_instance( [ 545.227614] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 545.227614] env[62569]: with excutils.save_and_reraise_exception(): [ 545.227614] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.227614] env[62569]: self.force_reraise() [ 545.227614] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.227614] env[62569]: raise self.value [ 545.227614] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 545.227614] env[62569]: updated_port = self._update_port( [ 545.227614] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.227614] env[62569]: _ensure_no_port_binding_failure(port) [ 545.227614] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.227614] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 545.228584] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 2a118c7a-8fdc-4948-9ec7-facf3bf84bd8, please check neutron logs for more information. [ 545.228584] env[62569]: Removing descriptor: 17 [ 545.228584] env[62569]: ERROR nova.compute.manager [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2a118c7a-8fdc-4948-9ec7-facf3bf84bd8, please check neutron logs for more information. [ 545.228584] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Traceback (most recent call last): [ 545.228584] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 545.228584] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] yield resources [ 545.228584] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 545.228584] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] self.driver.spawn(context, instance, image_meta, [ 545.228584] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 545.228584] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 545.228584] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 545.228584] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] vm_ref = self.build_virtual_machine(instance, [ 545.229369] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 545.229369] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] vif_infos = vmwarevif.get_vif_info(self._session, [ 545.229369] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 545.229369] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] for vif in network_info: [ 545.229369] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 545.229369] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] return self._sync_wrapper(fn, *args, **kwargs) [ 545.229369] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 545.229369] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] self.wait() [ 545.229369] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 545.229369] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] self[:] = self._gt.wait() [ 545.229369] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 545.229369] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] return self._exit_event.wait() [ 545.229369] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 545.229844] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] result = hub.switch() [ 545.229844] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 545.229844] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] return self.greenlet.switch() [ 545.229844] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.229844] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] result = function(*args, **kwargs) [ 545.229844] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 545.229844] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] return func(*args, **kwargs) [ 545.229844] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 545.229844] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] raise e [ 545.229844] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 545.229844] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] nwinfo = self.network_api.allocate_for_instance( [ 545.229844] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 545.229844] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] created_port_ids = self._update_ports_for_instance( [ 545.234828] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 545.234828] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] with excutils.save_and_reraise_exception(): [ 545.234828] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.234828] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] self.force_reraise() [ 545.234828] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.234828] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] raise self.value [ 545.234828] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 545.234828] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] updated_port = self._update_port( [ 545.234828] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.234828] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] _ensure_no_port_binding_failure(port) [ 545.234828] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.234828] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] raise exception.PortBindingFailed(port_id=port['id']) [ 545.235284] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] nova.exception.PortBindingFailed: Binding failed for port 2a118c7a-8fdc-4948-9ec7-facf3bf84bd8, please check neutron logs for more information. [ 545.235284] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] [ 545.235284] env[62569]: INFO nova.compute.manager [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Terminating instance [ 545.263825] env[62569]: DEBUG nova.compute.manager [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 545.479720] env[62569]: DEBUG nova.network.neutron [req-9c7f350a-5520-4d61-ba26-2bba5c4fc426 req-fcf72db7-1ea7-4699-867b-879868d0ad2c service nova] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.673368] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249823, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.102697} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.673975] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 545.674178] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 545.674374] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 545.740298] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Acquiring lock "refresh_cache-9fa91603-f812-4d93-bef1-58e455f2c2e8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.786426] env[62569]: DEBUG nova.network.neutron [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.792884] env[62569]: DEBUG oslo_concurrency.lockutils [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.835337] env[62569]: DEBUG nova.network.neutron [req-9c7f350a-5520-4d61-ba26-2bba5c4fc426 req-fcf72db7-1ea7-4699-867b-879868d0ad2c service nova] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.840068] env[62569]: DEBUG nova.network.neutron [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Successfully created port: 6ed7b301-12d9-4c35-880f-58425e991502 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 546.090099] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c875ab-6c59-4a56-9444-316f60ca7e53 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.100366] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d6ac7e-0a57-4e95-837a-f21be7b495b7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.136998] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2099dac-b397-4b23-b8cb-d74ac416dca9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.145764] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96f66293-58bf-4c9c-bcf2-166e97da9ca5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.160311] env[62569]: DEBUG nova.compute.provider_tree [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 546.296802] env[62569]: DEBUG nova.network.neutron [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.338260] env[62569]: DEBUG oslo_concurrency.lockutils [req-9c7f350a-5520-4d61-ba26-2bba5c4fc426 req-fcf72db7-1ea7-4699-867b-879868d0ad2c service nova] Releasing lock "refresh_cache-9fa91603-f812-4d93-bef1-58e455f2c2e8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.338657] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Acquired lock "refresh_cache-9fa91603-f812-4d93-bef1-58e455f2c2e8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.338835] env[62569]: DEBUG nova.network.neutron [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 546.663969] env[62569]: DEBUG nova.scheduler.client.report [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 546.713935] env[62569]: DEBUG nova.virt.hardware [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 546.714200] env[62569]: DEBUG nova.virt.hardware [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 546.714360] env[62569]: DEBUG nova.virt.hardware [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 546.714538] env[62569]: DEBUG nova.virt.hardware [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 546.714682] env[62569]: DEBUG nova.virt.hardware [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 546.714827] env[62569]: DEBUG nova.virt.hardware [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 546.715051] env[62569]: DEBUG nova.virt.hardware [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 546.715222] env[62569]: DEBUG nova.virt.hardware [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 546.715391] env[62569]: DEBUG nova.virt.hardware [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 546.715792] env[62569]: DEBUG nova.virt.hardware [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 546.715792] env[62569]: DEBUG nova.virt.hardware [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 546.716598] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef4e247c-e515-4be1-b51a-1dace553a937 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.725450] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67c77787-98b7-451b-827c-527326829d33 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.744686] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Instance VIF info [] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 546.753358] env[62569]: DEBUG oslo.service.loopingcall [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 546.753702] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 546.753896] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a72c6b17-7047-4e63-85f1-1a416efe3536 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.775692] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 546.775692] env[62569]: value = "task-1249824" [ 546.775692] env[62569]: _type = "Task" [ 546.775692] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.784675] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249824, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.800513] env[62569]: DEBUG oslo_concurrency.lockutils [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Releasing lock "refresh_cache-597c9673-bbd4-4dd0-829a-c05f8667c951" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.800751] env[62569]: DEBUG nova.compute.manager [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 546.801054] env[62569]: DEBUG nova.compute.manager [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 546.801794] env[62569]: DEBUG nova.network.neutron [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 546.848607] env[62569]: DEBUG nova.network.neutron [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 546.892139] env[62569]: DEBUG nova.network.neutron [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 547.170202] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.021s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.173310] env[62569]: ERROR nova.compute.manager [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 247970bb-e8a0-4e10-a07a-04e6c4ced856, please check neutron logs for more information. [ 547.173310] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Traceback (most recent call last): [ 547.173310] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 547.173310] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] self.driver.spawn(context, instance, image_meta, [ 547.173310] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 547.173310] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 547.173310] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 547.173310] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] vm_ref = self.build_virtual_machine(instance, [ 547.173310] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 547.173310] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] vif_infos = vmwarevif.get_vif_info(self._session, [ 547.173310] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 547.173706] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] for vif in network_info: [ 547.173706] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 547.173706] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] return self._sync_wrapper(fn, *args, **kwargs) [ 547.173706] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 547.173706] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] self.wait() [ 547.173706] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 547.173706] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] self[:] = self._gt.wait() [ 547.173706] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 547.173706] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] return self._exit_event.wait() [ 547.173706] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 547.173706] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] result = hub.switch() [ 547.173706] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 547.173706] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] return self.greenlet.switch() [ 547.174136] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.174136] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] result = function(*args, **kwargs) [ 547.174136] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 547.174136] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] return func(*args, **kwargs) [ 547.174136] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 547.174136] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] raise e [ 547.174136] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 547.174136] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] nwinfo = self.network_api.allocate_for_instance( [ 547.174136] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.174136] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] created_port_ids = self._update_ports_for_instance( [ 547.174136] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.174136] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] with excutils.save_and_reraise_exception(): [ 547.174136] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.174563] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] self.force_reraise() [ 547.174563] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.174563] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] raise self.value [ 547.174563] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.174563] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] updated_port = self._update_port( [ 547.174563] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.174563] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] _ensure_no_port_binding_failure(port) [ 547.174563] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.174563] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] raise exception.PortBindingFailed(port_id=port['id']) [ 547.174563] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] nova.exception.PortBindingFailed: Binding failed for port 247970bb-e8a0-4e10-a07a-04e6c4ced856, please check neutron logs for more information. [ 547.174563] env[62569]: ERROR nova.compute.manager [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] [ 547.174903] env[62569]: DEBUG nova.compute.utils [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Binding failed for port 247970bb-e8a0-4e10-a07a-04e6c4ced856, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 547.176024] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.661s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.177957] env[62569]: INFO nova.compute.claims [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 547.181367] env[62569]: DEBUG nova.compute.manager [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Build of instance c542f5dd-10d8-4959-a3ab-eab9b8f4525a was re-scheduled: Binding failed for port 247970bb-e8a0-4e10-a07a-04e6c4ced856, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 547.181840] env[62569]: DEBUG nova.compute.manager [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 547.182455] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Acquiring lock "refresh_cache-c542f5dd-10d8-4959-a3ab-eab9b8f4525a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.182623] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Acquired lock "refresh_cache-c542f5dd-10d8-4959-a3ab-eab9b8f4525a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.183011] env[62569]: DEBUG nova.network.neutron [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 547.271669] env[62569]: DEBUG nova.network.neutron [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.289678] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249824, 'name': CreateVM_Task, 'duration_secs': 0.314365} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.289908] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 547.290373] env[62569]: DEBUG oslo_concurrency.lockutils [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.293307] env[62569]: DEBUG oslo_concurrency.lockutils [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.293307] env[62569]: DEBUG oslo_concurrency.lockutils [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 547.293307] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34a9a400-fcb6-4d76-880a-91c967ea83dd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.299862] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Waiting for the task: (returnval){ [ 547.299862] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5207faf8-491e-bbc8-acdc-2d11aad63eec" [ 547.299862] env[62569]: _type = "Task" [ 547.299862] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.306839] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5207faf8-491e-bbc8-acdc-2d11aad63eec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.351405] env[62569]: DEBUG nova.network.neutron [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.743107] env[62569]: DEBUG nova.network.neutron [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 547.776450] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Releasing lock "refresh_cache-9fa91603-f812-4d93-bef1-58e455f2c2e8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.776450] env[62569]: DEBUG nova.compute.manager [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 547.776985] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 547.777359] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f1fcc3c1-d6ba-4b1d-a22e-671739d1fcef {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.789211] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de741353-e03b-4a1c-b940-0538c0bcf9a1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.818858] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5207faf8-491e-bbc8-acdc-2d11aad63eec, 'name': SearchDatastore_Task, 'duration_secs': 0.009212} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.823952] env[62569]: DEBUG oslo_concurrency.lockutils [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.824306] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 547.824482] env[62569]: DEBUG oslo_concurrency.lockutils [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.824628] env[62569]: DEBUG oslo_concurrency.lockutils [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.825516] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 547.825516] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9fa91603-f812-4d93-bef1-58e455f2c2e8 could not be found. [ 547.825823] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 547.825823] env[62569]: INFO nova.compute.manager [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Took 0.05 seconds to destroy the instance on the hypervisor. [ 547.829900] env[62569]: DEBUG oslo.service.loopingcall [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 547.829900] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-02f3ad7b-bd41-4eb8-bb19-5730edc6dfcf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.829900] env[62569]: DEBUG nova.compute.manager [-] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 547.829900] env[62569]: DEBUG nova.network.neutron [-] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 547.837734] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 547.837734] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 547.838091] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1602e235-3d11-47ac-bdfc-652b20e4bf32 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.847513] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Waiting for the task: (returnval){ [ 547.847513] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52615ce1-657c-e6c7-d620-e17f7079951c" [ 547.847513] env[62569]: _type = "Task" [ 547.847513] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.855873] env[62569]: INFO nova.compute.manager [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] Took 1.05 seconds to deallocate network for instance. [ 547.858580] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52615ce1-657c-e6c7-d620-e17f7079951c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.877812] env[62569]: DEBUG nova.network.neutron [-] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.200391] env[62569]: DEBUG nova.network.neutron [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.266124] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Acquiring lock "94818059-cf10-4dca-a711-fd5809583ce9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.266124] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Lock "94818059-cf10-4dca-a711-fd5809583ce9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.380930] env[62569]: DEBUG nova.network.neutron [-] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.382279] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52615ce1-657c-e6c7-d620-e17f7079951c, 'name': SearchDatastore_Task, 'duration_secs': 0.012546} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.388689] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b83f7470-7533-4041-a472-0124d53056c2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.400498] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Waiting for the task: (returnval){ [ 548.400498] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52177f54-a48b-a788-53e1-4fb0b590fd7d" [ 548.400498] env[62569]: _type = "Task" [ 548.400498] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.414200] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52177f54-a48b-a788-53e1-4fb0b590fd7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.581017] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d8c3e6a-15d7-4f07-ac13-d609888527ed {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.589070] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b911b369-f73f-4a8b-8cdb-59d70084e90e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.627458] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-278490b7-db07-466c-a0c1-f58f9ad291c5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.635585] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d663a54-9640-4b76-95b6-a93ab214306d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.654575] env[62569]: DEBUG nova.compute.provider_tree [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 548.706284] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Releasing lock "refresh_cache-c542f5dd-10d8-4959-a3ab-eab9b8f4525a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.706534] env[62569]: DEBUG nova.compute.manager [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 548.706715] env[62569]: DEBUG nova.compute.manager [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 548.706878] env[62569]: DEBUG nova.network.neutron [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 548.742168] env[62569]: DEBUG nova.network.neutron [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.884586] env[62569]: INFO nova.compute.manager [-] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Took 1.06 seconds to deallocate network for instance. [ 548.886920] env[62569]: DEBUG nova.compute.claims [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 548.887335] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.919509] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52177f54-a48b-a788-53e1-4fb0b590fd7d, 'name': SearchDatastore_Task, 'duration_secs': 0.014213} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.921192] env[62569]: DEBUG oslo_concurrency.lockutils [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.923092] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 6e35fe1c-05bc-4dcc-b7ec-e9998a647739/6e35fe1c-05bc-4dcc-b7ec-e9998a647739.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 548.923930] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f14a2922-8b82-4174-9d77-09f02b69252a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.929078] env[62569]: INFO nova.scheduler.client.report [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Deleted allocations for instance 597c9673-bbd4-4dd0-829a-c05f8667c951 [ 548.945547] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Waiting for the task: (returnval){ [ 548.945547] env[62569]: value = "task-1249825" [ 548.945547] env[62569]: _type = "Task" [ 548.945547] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.958301] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249825, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.157636] env[62569]: DEBUG nova.scheduler.client.report [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 549.208032] env[62569]: DEBUG nova.compute.manager [req-6f09a8b5-41a3-4809-8408-ba37f191ca60 req-5501b997-2d74-4afb-a112-d0048840a2b4 service nova] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Received event network-vif-deleted-2a118c7a-8fdc-4948-9ec7-facf3bf84bd8 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 549.248673] env[62569]: DEBUG nova.network.neutron [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.361878] env[62569]: ERROR nova.compute.manager [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6ed7b301-12d9-4c35-880f-58425e991502, please check neutron logs for more information. [ 549.361878] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 549.361878] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 549.361878] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 549.361878] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 549.361878] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 549.361878] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 549.361878] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 549.361878] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.361878] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 549.361878] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.361878] env[62569]: ERROR nova.compute.manager raise self.value [ 549.361878] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 549.361878] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 549.361878] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.361878] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 549.362442] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.362442] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 549.362442] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6ed7b301-12d9-4c35-880f-58425e991502, please check neutron logs for more information. [ 549.362442] env[62569]: ERROR nova.compute.manager [ 549.362442] env[62569]: Traceback (most recent call last): [ 549.362442] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 549.362442] env[62569]: listener.cb(fileno) [ 549.362442] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.362442] env[62569]: result = function(*args, **kwargs) [ 549.362442] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 549.362442] env[62569]: return func(*args, **kwargs) [ 549.362442] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 549.362442] env[62569]: raise e [ 549.362442] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 549.362442] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 549.362442] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 549.362442] env[62569]: created_port_ids = self._update_ports_for_instance( [ 549.362442] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 549.362442] env[62569]: with excutils.save_and_reraise_exception(): [ 549.362442] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.362442] env[62569]: self.force_reraise() [ 549.362442] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.362442] env[62569]: raise self.value [ 549.362442] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 549.362442] env[62569]: updated_port = self._update_port( [ 549.362442] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.362442] env[62569]: _ensure_no_port_binding_failure(port) [ 549.362442] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.362442] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 549.363357] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 6ed7b301-12d9-4c35-880f-58425e991502, please check neutron logs for more information. [ 549.363357] env[62569]: Removing descriptor: 14 [ 549.364762] env[62569]: ERROR nova.compute.manager [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6ed7b301-12d9-4c35-880f-58425e991502, please check neutron logs for more information. [ 549.364762] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Traceback (most recent call last): [ 549.364762] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 549.364762] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] yield resources [ 549.364762] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 549.364762] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] self.driver.spawn(context, instance, image_meta, [ 549.364762] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 549.364762] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 549.364762] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 549.364762] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] vm_ref = self.build_virtual_machine(instance, [ 549.364762] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 549.365251] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] vif_infos = vmwarevif.get_vif_info(self._session, [ 549.365251] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 549.365251] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] for vif in network_info: [ 549.365251] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 549.365251] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] return self._sync_wrapper(fn, *args, **kwargs) [ 549.365251] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 549.365251] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] self.wait() [ 549.365251] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 549.365251] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] self[:] = self._gt.wait() [ 549.365251] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 549.365251] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] return self._exit_event.wait() [ 549.365251] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 549.365251] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] result = hub.switch() [ 549.365617] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 549.365617] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] return self.greenlet.switch() [ 549.365617] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.365617] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] result = function(*args, **kwargs) [ 549.365617] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 549.365617] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] return func(*args, **kwargs) [ 549.365617] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 549.365617] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] raise e [ 549.365617] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 549.365617] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] nwinfo = self.network_api.allocate_for_instance( [ 549.365617] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 549.365617] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] created_port_ids = self._update_ports_for_instance( [ 549.365617] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 549.365990] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] with excutils.save_and_reraise_exception(): [ 549.365990] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.365990] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] self.force_reraise() [ 549.365990] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.365990] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] raise self.value [ 549.365990] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 549.365990] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] updated_port = self._update_port( [ 549.365990] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.365990] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] _ensure_no_port_binding_failure(port) [ 549.365990] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.365990] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] raise exception.PortBindingFailed(port_id=port['id']) [ 549.365990] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] nova.exception.PortBindingFailed: Binding failed for port 6ed7b301-12d9-4c35-880f-58425e991502, please check neutron logs for more information. [ 549.365990] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] [ 549.366365] env[62569]: INFO nova.compute.manager [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Terminating instance [ 549.439030] env[62569]: DEBUG oslo_concurrency.lockutils [None req-45f46cdc-f494-4e59-a182-1058124dfc53 tempest-VolumesAssistedSnapshotsTest-1144329263 tempest-VolumesAssistedSnapshotsTest-1144329263-project-member] Lock "597c9673-bbd4-4dd0-829a-c05f8667c951" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.802s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.443241] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "597c9673-bbd4-4dd0-829a-c05f8667c951" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 24.930s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.443306] env[62569]: INFO nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 597c9673-bbd4-4dd0-829a-c05f8667c951] During sync_power_state the instance has a pending task (spawning). Skip. [ 549.444064] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "597c9673-bbd4-4dd0-829a-c05f8667c951" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.462856] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249825, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485349} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.462856] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 6e35fe1c-05bc-4dcc-b7ec-e9998a647739/6e35fe1c-05bc-4dcc-b7ec-e9998a647739.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 549.463017] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 549.463257] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f64f7396-cfd3-4100-b9d3-391cb32ff309 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.471026] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Waiting for the task: (returnval){ [ 549.471026] env[62569]: value = "task-1249826" [ 549.471026] env[62569]: _type = "Task" [ 549.471026] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.482707] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249826, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.669592] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.494s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.670235] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 549.673336] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 10.978s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.698764] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Acquiring lock "c29ad19e-91a5-4c3d-912b-bd7be4c42016" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.698989] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Lock "c29ad19e-91a5-4c3d-912b-bd7be4c42016" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.756634] env[62569]: INFO nova.compute.manager [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] Took 1.05 seconds to deallocate network for instance. [ 549.871427] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Acquiring lock "refresh_cache-5b1117bd-6210-4df2-b828-f76951752f2f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.871427] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Acquired lock "refresh_cache-5b1117bd-6210-4df2-b828-f76951752f2f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.871564] env[62569]: DEBUG nova.network.neutron [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 549.952737] env[62569]: DEBUG nova.compute.manager [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 549.987023] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249826, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067961} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.987876] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 549.989659] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc7979d3-f8e8-4494-80cc-8788aab5ee2c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.015082] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Reconfiguring VM instance instance-0000000c to attach disk [datastore1] 6e35fe1c-05bc-4dcc-b7ec-e9998a647739/6e35fe1c-05bc-4dcc-b7ec-e9998a647739.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 550.015082] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-992a40d7-400a-4643-a7b8-c5685ec4c866 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.034795] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Waiting for the task: (returnval){ [ 550.034795] env[62569]: value = "task-1249827" [ 550.034795] env[62569]: _type = "Task" [ 550.034795] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.043158] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249827, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.188266] env[62569]: DEBUG nova.compute.utils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 550.193020] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 550.193020] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 550.306507] env[62569]: DEBUG nova.policy [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '599a3e09a44d4b6a980e34ba73f9d6e5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff4d4d8e74664c3a9fb2ab8ca80ba126', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 550.418876] env[62569]: DEBUG nova.network.neutron [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 550.487251] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.550041] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249827, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.682483] env[62569]: DEBUG nova.network.neutron [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.697362] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 550.729651] env[62569]: WARNING nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 550.806168] env[62569]: INFO nova.scheduler.client.report [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Deleted allocations for instance c542f5dd-10d8-4959-a3ab-eab9b8f4525a [ 551.050067] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249827, 'name': ReconfigVM_Task, 'duration_secs': 0.81198} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.050439] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Reconfigured VM instance instance-0000000c to attach disk [datastore1] 6e35fe1c-05bc-4dcc-b7ec-e9998a647739/6e35fe1c-05bc-4dcc-b7ec-e9998a647739.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 551.054574] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-981519af-38d2-4b58-bf44-98381472e817 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.061811] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Waiting for the task: (returnval){ [ 551.061811] env[62569]: value = "task-1249828" [ 551.061811] env[62569]: _type = "Task" [ 551.061811] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.074892] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249828, 'name': Rename_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.193019] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Releasing lock "refresh_cache-5b1117bd-6210-4df2-b828-f76951752f2f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 551.193019] env[62569]: DEBUG nova.compute.manager [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 551.193019] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 551.193019] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-697a1981-46e0-41b9-87c8-950f1cb406bb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.204014] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b1ed8f-0806-449a-9ef5-8e7336919966 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.228074] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5b1117bd-6210-4df2-b828-f76951752f2f could not be found. [ 551.228074] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 551.228074] env[62569]: INFO nova.compute.manager [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 551.228238] env[62569]: DEBUG oslo.service.loopingcall [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 551.231037] env[62569]: DEBUG nova.compute.manager [-] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 551.231037] env[62569]: DEBUG nova.network.neutron [-] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 551.235039] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance c542f5dd-10d8-4959-a3ab-eab9b8f4525a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 551.235039] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance ba23eea7-422c-4e5d-b95e-f95589e41550 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 551.235039] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance e8b8b029-b20a-4af3-a5dc-1b032ce46f8d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 551.236900] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 6e35fe1c-05bc-4dcc-b7ec-e9998a647739 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 551.236900] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 9fa91603-f812-4d93-bef1-58e455f2c2e8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 551.236900] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 5b1117bd-6210-4df2-b828-f76951752f2f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 551.236900] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 232a126f-37ec-4aea-a602-eb3a13d1d70a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 551.292216] env[62569]: DEBUG nova.network.neutron [-] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 551.307800] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Successfully created port: dcaa61ad-1005-48a6-bbe4-157575908f59 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 551.318010] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca232298-7846-4b91-befb-b67f42d3f311 tempest-ServersWithSpecificFlavorTestJSON-1434969159 tempest-ServersWithSpecificFlavorTestJSON-1434969159-project-member] Lock "c542f5dd-10d8-4959-a3ab-eab9b8f4525a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.797s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.321837] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "c542f5dd-10d8-4959-a3ab-eab9b8f4525a" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 26.807s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.321837] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2cc12fc9-f382-438e-8c8c-103e366b2842 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.330937] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d54e8da4-a8f3-485b-b9b5-2b1daec8ac38 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.574994] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249828, 'name': Rename_Task, 'duration_secs': 0.137611} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.575702] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 551.575702] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1ec6d6fe-b027-47da-bce6-224102c61bc4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.584223] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Waiting for the task: (returnval){ [ 551.584223] env[62569]: value = "task-1249829" [ 551.584223] env[62569]: _type = "Task" [ 551.584223] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.602539] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249829, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.719237] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 551.740369] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 11c53593-cf14-4e46-b546-c90a3184028e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 551.750400] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 551.751241] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 551.751241] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 551.751241] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 551.751241] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 551.751567] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 551.751855] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 551.752400] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 551.752400] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 551.752400] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 551.752920] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 551.753841] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41b7405e-98e9-4e27-a4f9-172b848b23be {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.764775] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fce2287-c78e-4b20-8cb6-dcc0840ec71c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.795578] env[62569]: DEBUG nova.network.neutron [-] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.822180] env[62569]: DEBUG nova.compute.manager [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 551.858875] env[62569]: INFO nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: c542f5dd-10d8-4959-a3ab-eab9b8f4525a] During the sync_power process the instance has moved from host None to host cpu-1 [ 551.859083] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "c542f5dd-10d8-4959-a3ab-eab9b8f4525a" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.539s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.096502] env[62569]: DEBUG oslo_vmware.api [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249829, 'name': PowerOnVM_Task, 'duration_secs': 0.426771} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 552.097539] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 552.098882] env[62569]: DEBUG nova.compute.manager [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 552.101017] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4e26e23-f065-43b2-8326-0f611a9d843a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.117893] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "418c648a-06be-4e97-b138-45c65addf5cf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.118280] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "418c648a-06be-4e97-b138-45c65addf5cf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.165630] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "92666c7b-7a1d-4d49-931f-e84cd2486871" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.165630] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "92666c7b-7a1d-4d49-931f-e84cd2486871" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.243659] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance a9224414-6d20-4d5f-be5d-2b893117f4b8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 552.302919] env[62569]: INFO nova.compute.manager [-] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Took 1.07 seconds to deallocate network for instance. [ 552.306038] env[62569]: DEBUG nova.compute.claims [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 552.306299] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.360467] env[62569]: DEBUG oslo_concurrency.lockutils [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.507772] env[62569]: DEBUG nova.compute.manager [req-c5bb0fd4-7473-40cb-97b5-8a31bae76654 req-9c7d76c6-0a13-4ff2-9dcc-15b90d738cc4 service nova] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Received event network-changed-6ed7b301-12d9-4c35-880f-58425e991502 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 552.507772] env[62569]: DEBUG nova.compute.manager [req-c5bb0fd4-7473-40cb-97b5-8a31bae76654 req-9c7d76c6-0a13-4ff2-9dcc-15b90d738cc4 service nova] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Refreshing instance network info cache due to event network-changed-6ed7b301-12d9-4c35-880f-58425e991502. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 552.507772] env[62569]: DEBUG oslo_concurrency.lockutils [req-c5bb0fd4-7473-40cb-97b5-8a31bae76654 req-9c7d76c6-0a13-4ff2-9dcc-15b90d738cc4 service nova] Acquiring lock "refresh_cache-5b1117bd-6210-4df2-b828-f76951752f2f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.507772] env[62569]: DEBUG oslo_concurrency.lockutils [req-c5bb0fd4-7473-40cb-97b5-8a31bae76654 req-9c7d76c6-0a13-4ff2-9dcc-15b90d738cc4 service nova] Acquired lock "refresh_cache-5b1117bd-6210-4df2-b828-f76951752f2f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.508634] env[62569]: DEBUG nova.network.neutron [req-c5bb0fd4-7473-40cb-97b5-8a31bae76654 req-9c7d76c6-0a13-4ff2-9dcc-15b90d738cc4 service nova] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Refreshing network info cache for port 6ed7b301-12d9-4c35-880f-58425e991502 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 552.630479] env[62569]: DEBUG oslo_concurrency.lockutils [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.675806] env[62569]: DEBUG oslo_concurrency.lockutils [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Acquiring lock "21cde60b-2764-4ff7-ab5e-a17fb672351d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.676091] env[62569]: DEBUG oslo_concurrency.lockutils [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Lock "21cde60b-2764-4ff7-ab5e-a17fb672351d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.747744] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 49fdf371-5278-442b-b8cc-9a0fc22947b4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 553.050364] env[62569]: DEBUG nova.network.neutron [req-c5bb0fd4-7473-40cb-97b5-8a31bae76654 req-9c7d76c6-0a13-4ff2-9dcc-15b90d738cc4 service nova] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.252568] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance ef52ac4a-fa98-4af8-b08a-b34cae526850 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 553.458633] env[62569]: DEBUG nova.network.neutron [req-c5bb0fd4-7473-40cb-97b5-8a31bae76654 req-9c7d76c6-0a13-4ff2-9dcc-15b90d738cc4 service nova] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.755953] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 0dc705ca-6fd1-4844-9c2d-34f2c099ea67 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 553.903298] env[62569]: INFO nova.compute.manager [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Rebuilding instance [ 553.956775] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Acquiring lock "d53a6920-d70c-40a6-a059-162a5f876b29" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.957046] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Lock "d53a6920-d70c-40a6-a059-162a5f876b29" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.963093] env[62569]: DEBUG oslo_concurrency.lockutils [req-c5bb0fd4-7473-40cb-97b5-8a31bae76654 req-9c7d76c6-0a13-4ff2-9dcc-15b90d738cc4 service nova] Releasing lock "refresh_cache-5b1117bd-6210-4df2-b828-f76951752f2f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.963320] env[62569]: DEBUG nova.compute.manager [req-c5bb0fd4-7473-40cb-97b5-8a31bae76654 req-9c7d76c6-0a13-4ff2-9dcc-15b90d738cc4 service nova] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Received event network-vif-deleted-6ed7b301-12d9-4c35-880f-58425e991502 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 553.968271] env[62569]: DEBUG nova.compute.manager [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 553.971451] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbeeb7c5-3289-44f4-802d-8b491a577c4a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.260682] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 21af21da-87b1-4a84-8544-f1a88494eb08 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 554.745071] env[62569]: ERROR nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dcaa61ad-1005-48a6-bbe4-157575908f59, please check neutron logs for more information. [ 554.745071] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 554.745071] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 554.745071] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 554.745071] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 554.745071] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 554.745071] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 554.745071] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 554.745071] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.745071] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 554.745071] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.745071] env[62569]: ERROR nova.compute.manager raise self.value [ 554.745071] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 554.745071] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 554.745071] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.745071] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 554.746222] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.746222] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 554.746222] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dcaa61ad-1005-48a6-bbe4-157575908f59, please check neutron logs for more information. [ 554.746222] env[62569]: ERROR nova.compute.manager [ 554.746222] env[62569]: Traceback (most recent call last): [ 554.746222] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 554.746222] env[62569]: listener.cb(fileno) [ 554.746222] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 554.746222] env[62569]: result = function(*args, **kwargs) [ 554.746222] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 554.746222] env[62569]: return func(*args, **kwargs) [ 554.746222] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 554.746222] env[62569]: raise e [ 554.746222] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 554.746222] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 554.746222] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 554.746222] env[62569]: created_port_ids = self._update_ports_for_instance( [ 554.746222] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 554.746222] env[62569]: with excutils.save_and_reraise_exception(): [ 554.746222] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.746222] env[62569]: self.force_reraise() [ 554.746222] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.746222] env[62569]: raise self.value [ 554.746222] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 554.746222] env[62569]: updated_port = self._update_port( [ 554.746222] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.746222] env[62569]: _ensure_no_port_binding_failure(port) [ 554.746222] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.746222] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 554.747123] env[62569]: nova.exception.PortBindingFailed: Binding failed for port dcaa61ad-1005-48a6-bbe4-157575908f59, please check neutron logs for more information. [ 554.747123] env[62569]: Removing descriptor: 14 [ 554.747123] env[62569]: ERROR nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dcaa61ad-1005-48a6-bbe4-157575908f59, please check neutron logs for more information. [ 554.747123] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Traceback (most recent call last): [ 554.747123] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 554.747123] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] yield resources [ 554.747123] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 554.747123] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] self.driver.spawn(context, instance, image_meta, [ 554.747123] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 554.747123] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 554.747123] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 554.747123] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] vm_ref = self.build_virtual_machine(instance, [ 554.747549] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 554.747549] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] vif_infos = vmwarevif.get_vif_info(self._session, [ 554.747549] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 554.747549] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] for vif in network_info: [ 554.747549] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 554.747549] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] return self._sync_wrapper(fn, *args, **kwargs) [ 554.747549] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 554.747549] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] self.wait() [ 554.747549] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 554.747549] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] self[:] = self._gt.wait() [ 554.747549] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 554.747549] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] return self._exit_event.wait() [ 554.747549] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 554.747981] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] result = hub.switch() [ 554.747981] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 554.747981] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] return self.greenlet.switch() [ 554.747981] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 554.747981] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] result = function(*args, **kwargs) [ 554.747981] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 554.747981] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] return func(*args, **kwargs) [ 554.747981] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 554.747981] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] raise e [ 554.747981] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 554.747981] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] nwinfo = self.network_api.allocate_for_instance( [ 554.747981] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 554.747981] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] created_port_ids = self._update_ports_for_instance( [ 554.748849] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 554.748849] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] with excutils.save_and_reraise_exception(): [ 554.748849] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.748849] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] self.force_reraise() [ 554.748849] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.748849] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] raise self.value [ 554.748849] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 554.748849] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] updated_port = self._update_port( [ 554.748849] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.748849] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] _ensure_no_port_binding_failure(port) [ 554.748849] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.748849] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] raise exception.PortBindingFailed(port_id=port['id']) [ 554.749364] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] nova.exception.PortBindingFailed: Binding failed for port dcaa61ad-1005-48a6-bbe4-157575908f59, please check neutron logs for more information. [ 554.749364] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] [ 554.749364] env[62569]: INFO nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Terminating instance [ 554.764135] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance c6292246-847f-4466-909b-6139401943af has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 554.991403] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 554.991726] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-829c03c1-5402-451e-baed-428f3ca19c62 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.000222] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Waiting for the task: (returnval){ [ 555.000222] env[62569]: value = "task-1249830" [ 555.000222] env[62569]: _type = "Task" [ 555.000222] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.010808] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Task: {'id': task-1249830, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.256745] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquiring lock "refresh_cache-232a126f-37ec-4aea-a602-eb3a13d1d70a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.256936] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquired lock "refresh_cache-232a126f-37ec-4aea-a602-eb3a13d1d70a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.257180] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 555.267201] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 2a61b4cd-a15c-4064-9351-326dd777603b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 555.289382] env[62569]: DEBUG nova.compute.manager [req-142102c5-730b-4343-a4cf-f4739101f6f0 req-2e8ad29c-cb49-4c2a-857b-7450a7fb667e service nova] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Received event network-changed-dcaa61ad-1005-48a6-bbe4-157575908f59 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 555.289584] env[62569]: DEBUG nova.compute.manager [req-142102c5-730b-4343-a4cf-f4739101f6f0 req-2e8ad29c-cb49-4c2a-857b-7450a7fb667e service nova] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Refreshing instance network info cache due to event network-changed-dcaa61ad-1005-48a6-bbe4-157575908f59. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 555.289764] env[62569]: DEBUG oslo_concurrency.lockutils [req-142102c5-730b-4343-a4cf-f4739101f6f0 req-2e8ad29c-cb49-4c2a-857b-7450a7fb667e service nova] Acquiring lock "refresh_cache-232a126f-37ec-4aea-a602-eb3a13d1d70a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.511446] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Task: {'id': task-1249830, 'name': PowerOffVM_Task, 'duration_secs': 0.123247} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.511898] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 555.512261] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 555.513128] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c8ca7f3-f8cb-44a7-a2df-6419db767dd8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.520614] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 555.521192] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c1bfef32-c681-4656-8f4f-e274a65ed557 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.547267] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 555.547267] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 555.547267] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Deleting the datastore file [datastore1] 6e35fe1c-05bc-4dcc-b7ec-e9998a647739 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 555.547267] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-620c5c31-4ae0-4a8b-bfef-5569ca45e54c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.557018] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Waiting for the task: (returnval){ [ 555.557018] env[62569]: value = "task-1249832" [ 555.557018] env[62569]: _type = "Task" [ 555.557018] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.567492] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Task: {'id': task-1249832, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.775524] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 76bddf93-ca4f-4987-a446-b369801ebd93 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 555.785382] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 555.923710] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Acquiring lock "e53104ad-35c7-49cb-b8de-035006148c52" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.924436] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Lock "e53104ad-35c7-49cb-b8de-035006148c52" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.008649] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.065131] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Task: {'id': task-1249832, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.103876} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.065466] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 556.065798] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 556.066116] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 556.280379] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 4b9e4f73-4682-4170-b71e-1d3d1d860577 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 556.513940] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Releasing lock "refresh_cache-232a126f-37ec-4aea-a602-eb3a13d1d70a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.514415] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 556.514610] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 556.514925] env[62569]: DEBUG oslo_concurrency.lockutils [req-142102c5-730b-4343-a4cf-f4739101f6f0 req-2e8ad29c-cb49-4c2a-857b-7450a7fb667e service nova] Acquired lock "refresh_cache-232a126f-37ec-4aea-a602-eb3a13d1d70a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.515365] env[62569]: DEBUG nova.network.neutron [req-142102c5-730b-4343-a4cf-f4739101f6f0 req-2e8ad29c-cb49-4c2a-857b-7450a7fb667e service nova] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Refreshing network info cache for port dcaa61ad-1005-48a6-bbe4-157575908f59 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 556.516173] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-70c81410-0cb5-497c-949f-16878bebf41f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.522863] env[62569]: DEBUG oslo_concurrency.lockutils [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Acquiring lock "7cb248f2-2db2-4f62-b2d1-332488170c9b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.522863] env[62569]: DEBUG oslo_concurrency.lockutils [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Lock "7cb248f2-2db2-4f62-b2d1-332488170c9b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.528865] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a880aab0-c6d3-4862-aca1-7787c95191a4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.557619] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 232a126f-37ec-4aea-a602-eb3a13d1d70a could not be found. [ 556.557954] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 556.558041] env[62569]: INFO nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 556.558266] env[62569]: DEBUG oslo.service.loopingcall [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 556.559100] env[62569]: DEBUG nova.compute.manager [-] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 556.559100] env[62569]: DEBUG nova.network.neutron [-] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 556.605505] env[62569]: DEBUG nova.network.neutron [-] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.783913] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 898f69fe-737a-44cb-9c43-965f1f8685f7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 557.055772] env[62569]: DEBUG nova.network.neutron [req-142102c5-730b-4343-a4cf-f4739101f6f0 req-2e8ad29c-cb49-4c2a-857b-7450a7fb667e service nova] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 557.105539] env[62569]: DEBUG nova.network.neutron [-] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.124859] env[62569]: DEBUG nova.virt.hardware [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 557.125366] env[62569]: DEBUG nova.virt.hardware [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 557.128090] env[62569]: DEBUG nova.virt.hardware [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 557.128090] env[62569]: DEBUG nova.virt.hardware [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 557.128090] env[62569]: DEBUG nova.virt.hardware [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 557.128090] env[62569]: DEBUG nova.virt.hardware [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 557.128090] env[62569]: DEBUG nova.virt.hardware [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 557.128300] env[62569]: DEBUG nova.virt.hardware [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 557.128300] env[62569]: DEBUG nova.virt.hardware [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 557.128300] env[62569]: DEBUG nova.virt.hardware [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 557.128300] env[62569]: DEBUG nova.virt.hardware [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 557.129139] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7895704-ecfa-42f8-b440-faf3bf64697d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.146850] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d060c880-109b-4c6a-844b-449a428a5a54 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.165226] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Instance VIF info [] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 557.173725] env[62569]: DEBUG oslo.service.loopingcall [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 557.174452] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 557.174602] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2052cbfe-9629-4e1c-92d5-9e178764148e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.193351] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 557.193351] env[62569]: value = "task-1249833" [ 557.193351] env[62569]: _type = "Task" [ 557.193351] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.202357] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249833, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.291390] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 94818059-cf10-4dca-a711-fd5809583ce9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 557.394352] env[62569]: DEBUG nova.network.neutron [req-142102c5-730b-4343-a4cf-f4739101f6f0 req-2e8ad29c-cb49-4c2a-857b-7450a7fb667e service nova] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.612496] env[62569]: INFO nova.compute.manager [-] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Took 1.05 seconds to deallocate network for instance. [ 557.614935] env[62569]: DEBUG nova.compute.claims [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 557.615209] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.709444] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249833, 'name': CreateVM_Task, 'duration_secs': 0.391224} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.709444] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 557.709444] env[62569]: DEBUG oslo_concurrency.lockutils [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.709444] env[62569]: DEBUG oslo_concurrency.lockutils [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.709444] env[62569]: DEBUG oslo_concurrency.lockutils [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 557.709444] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eaf51006-9336-41a8-bda5-3f32c6103db0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.713964] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Waiting for the task: (returnval){ [ 557.713964] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a19884-0085-4b55-3792-4e68c8c8f1b5" [ 557.713964] env[62569]: _type = "Task" [ 557.713964] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.722849] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a19884-0085-4b55-3792-4e68c8c8f1b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.798173] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance c29ad19e-91a5-4c3d-912b-bd7be4c42016 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 557.798173] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=62569) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 557.798173] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=62569) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 557.901562] env[62569]: DEBUG oslo_concurrency.lockutils [req-142102c5-730b-4343-a4cf-f4739101f6f0 req-2e8ad29c-cb49-4c2a-857b-7450a7fb667e service nova] Releasing lock "refresh_cache-232a126f-37ec-4aea-a602-eb3a13d1d70a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 558.081124] env[62569]: DEBUG nova.compute.manager [req-9fc7747c-5522-4b38-b9b3-6bafd34a0f4b req-a03709dd-7b4d-4b09-92af-b667de868128 service nova] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Received event network-vif-deleted-dcaa61ad-1005-48a6-bbe4-157575908f59 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 558.093790] env[62569]: DEBUG oslo_concurrency.lockutils [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquiring lock "695022e4-bb58-44d7-87e5-58bbd577ec7b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.094479] env[62569]: DEBUG oslo_concurrency.lockutils [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "695022e4-bb58-44d7-87e5-58bbd577ec7b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.227038] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a19884-0085-4b55-3792-4e68c8c8f1b5, 'name': SearchDatastore_Task, 'duration_secs': 0.009635} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.231704] env[62569]: DEBUG oslo_concurrency.lockutils [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 558.231704] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 558.231704] env[62569]: DEBUG oslo_concurrency.lockutils [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 558.231704] env[62569]: DEBUG oslo_concurrency.lockutils [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 558.232104] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 558.234270] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fe9bd5b0-cbf3-4aec-89c5-0a8598f985e1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.244458] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 558.244458] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 558.245328] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9aa9f762-2fd9-4496-b6a0-a89ea37d6ade {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.253280] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Waiting for the task: (returnval){ [ 558.253280] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5275fd35-f680-fd65-be00-418184c46edd" [ 558.253280] env[62569]: _type = "Task" [ 558.253280] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 558.261408] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5275fd35-f680-fd65-be00-418184c46edd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.293512] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba04816-57bb-442a-a0b4-9890990c36a5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.302127] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12f85147-72a1-4df8-ad86-32b0101a7d49 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.338084] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42b4761c-3d42-48b3-9edd-b73344faa24e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.347422] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90db07d1-d0d1-416c-9718-5c38243a5d95 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.360756] env[62569]: DEBUG nova.compute.provider_tree [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 558.770025] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5275fd35-f680-fd65-be00-418184c46edd, 'name': SearchDatastore_Task, 'duration_secs': 0.008973} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.770782] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c547d480-5a37-4849-a190-b6e53d3885a0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.778139] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Waiting for the task: (returnval){ [ 558.778139] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528a2efa-79ac-db66-0073-fb15ddc4ace1" [ 558.778139] env[62569]: _type = "Task" [ 558.778139] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 558.787718] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528a2efa-79ac-db66-0073-fb15ddc4ace1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.864334] env[62569]: DEBUG nova.scheduler.client.report [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 559.243435] env[62569]: DEBUG oslo_concurrency.lockutils [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "0eff7f21-17f8-4642-bfc4-a7a15bdfb41f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.244986] env[62569]: DEBUG oslo_concurrency.lockutils [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "0eff7f21-17f8-4642-bfc4-a7a15bdfb41f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.288548] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528a2efa-79ac-db66-0073-fb15ddc4ace1, 'name': SearchDatastore_Task, 'duration_secs': 0.01033} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.288876] env[62569]: DEBUG oslo_concurrency.lockutils [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.289182] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 6e35fe1c-05bc-4dcc-b7ec-e9998a647739/6e35fe1c-05bc-4dcc-b7ec-e9998a647739.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 559.289501] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5c0c5d8e-67b2-4f2c-bd64-d639a02cfb7f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.295738] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Waiting for the task: (returnval){ [ 559.295738] env[62569]: value = "task-1249834" [ 559.295738] env[62569]: _type = "Task" [ 559.295738] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.304576] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Task: {'id': task-1249834, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.371395] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62569) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 559.371395] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 9.698s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.374067] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.253s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.376959] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 559.377663] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Getting list of instances from cluster (obj){ [ 559.377663] env[62569]: value = "domain-c8" [ 559.377663] env[62569]: _type = "ClusterComputeResource" [ 559.377663] env[62569]: } {{(pid=62569) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 559.379337] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8639b416-2a46-4314-b7ce-03ba959cdc24 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.389890] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Got total of 1 instances {{(pid=62569) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 559.807897] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Task: {'id': task-1249834, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.479829} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.808144] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 6e35fe1c-05bc-4dcc-b7ec-e9998a647739/6e35fe1c-05bc-4dcc-b7ec-e9998a647739.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 559.808369] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 559.808835] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-28ac8847-ff80-4ce3-afcc-f7d3719be0b7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.818499] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Waiting for the task: (returnval){ [ 559.818499] env[62569]: value = "task-1249835" [ 559.818499] env[62569]: _type = "Task" [ 559.818499] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.834620] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Task: {'id': task-1249835, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.995374] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Acquiring lock "0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.995374] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Lock "0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.302459] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-908e91df-3886-4f8f-809c-620f72f88c55 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.310249] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb7b8b7-2651-4c10-8239-84b08b6cf666 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.346927] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb6c0474-ab15-4f1a-9a24-a2ff6f98e540 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.360639] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Task: {'id': task-1249835, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073687} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 560.361759] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bdf213a-9e7b-42f2-b5e6-9a8886bfe78d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.365628] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 560.367096] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10f78072-8d69-4e5d-a49d-9f6c70768511 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.386022] env[62569]: DEBUG nova.compute.provider_tree [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 560.400532] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Reconfiguring VM instance instance-0000000c to attach disk [datastore1] 6e35fe1c-05bc-4dcc-b7ec-e9998a647739/6e35fe1c-05bc-4dcc-b7ec-e9998a647739.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 560.401349] env[62569]: DEBUG nova.scheduler.client.report [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 560.404985] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d76b7cc-2603-4795-ad4d-4289298c2199 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.420056] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.048s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.420693] env[62569]: ERROR nova.compute.manager [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ae242046-3aec-42af-a3cc-1839937c8497, please check neutron logs for more information. [ 560.420693] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Traceback (most recent call last): [ 560.420693] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 560.420693] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] self.driver.spawn(context, instance, image_meta, [ 560.420693] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 560.420693] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 560.420693] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 560.420693] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] vm_ref = self.build_virtual_machine(instance, [ 560.420693] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 560.420693] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] vif_infos = vmwarevif.get_vif_info(self._session, [ 560.420693] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 560.421141] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] for vif in network_info: [ 560.421141] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 560.421141] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] return self._sync_wrapper(fn, *args, **kwargs) [ 560.421141] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 560.421141] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] self.wait() [ 560.421141] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 560.421141] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] self[:] = self._gt.wait() [ 560.421141] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 560.421141] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] return self._exit_event.wait() [ 560.421141] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 560.421141] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] result = hub.switch() [ 560.421141] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 560.421141] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] return self.greenlet.switch() [ 560.421591] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 560.421591] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] result = function(*args, **kwargs) [ 560.421591] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 560.421591] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] return func(*args, **kwargs) [ 560.421591] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 560.421591] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] raise e [ 560.421591] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 560.421591] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] nwinfo = self.network_api.allocate_for_instance( [ 560.421591] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 560.421591] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] created_port_ids = self._update_ports_for_instance( [ 560.421591] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 560.421591] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] with excutils.save_and_reraise_exception(): [ 560.421591] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.422044] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] self.force_reraise() [ 560.422044] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.422044] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] raise self.value [ 560.422044] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 560.422044] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] updated_port = self._update_port( [ 560.422044] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.422044] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] _ensure_no_port_binding_failure(port) [ 560.422044] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.422044] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] raise exception.PortBindingFailed(port_id=port['id']) [ 560.422044] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] nova.exception.PortBindingFailed: Binding failed for port ae242046-3aec-42af-a3cc-1839937c8497, please check neutron logs for more information. [ 560.422044] env[62569]: ERROR nova.compute.manager [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] [ 560.422424] env[62569]: DEBUG nova.compute.utils [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Binding failed for port ae242046-3aec-42af-a3cc-1839937c8497, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 560.422653] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.412s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.424327] env[62569]: INFO nova.compute.claims [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 560.427676] env[62569]: DEBUG nova.compute.manager [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Build of instance e8b8b029-b20a-4af3-a5dc-1b032ce46f8d was re-scheduled: Binding failed for port ae242046-3aec-42af-a3cc-1839937c8497, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 560.427982] env[62569]: DEBUG nova.compute.manager [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 560.428230] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Acquiring lock "refresh_cache-e8b8b029-b20a-4af3-a5dc-1b032ce46f8d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.428375] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Acquired lock "refresh_cache-e8b8b029-b20a-4af3-a5dc-1b032ce46f8d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.428530] env[62569]: DEBUG nova.network.neutron [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 560.435379] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Waiting for the task: (returnval){ [ 560.435379] env[62569]: value = "task-1249836" [ 560.435379] env[62569]: _type = "Task" [ 560.435379] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 560.444665] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Task: {'id': task-1249836, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.946390] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Task: {'id': task-1249836, 'name': ReconfigVM_Task, 'duration_secs': 0.304783} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 560.946606] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Reconfigured VM instance instance-0000000c to attach disk [datastore1] 6e35fe1c-05bc-4dcc-b7ec-e9998a647739/6e35fe1c-05bc-4dcc-b7ec-e9998a647739.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 560.947433] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a778ee96-2d51-46fd-b51c-412e38b00840 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.954115] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Waiting for the task: (returnval){ [ 560.954115] env[62569]: value = "task-1249837" [ 560.954115] env[62569]: _type = "Task" [ 560.954115] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 560.958053] env[62569]: DEBUG nova.network.neutron [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 560.969342] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Task: {'id': task-1249837, 'name': Rename_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.143284] env[62569]: DEBUG nova.network.neutron [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.472023] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Task: {'id': task-1249837, 'name': Rename_Task, 'duration_secs': 0.141432} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 561.472023] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 561.472158] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-693d6ee5-131a-4a50-a43f-b4e02517fb8b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.478837] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Waiting for the task: (returnval){ [ 561.478837] env[62569]: value = "task-1249838" [ 561.478837] env[62569]: _type = "Task" [ 561.478837] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.496045] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Task: {'id': task-1249838, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.650316] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Releasing lock "refresh_cache-e8b8b029-b20a-4af3-a5dc-1b032ce46f8d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.650625] env[62569]: DEBUG nova.compute.manager [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 561.651195] env[62569]: DEBUG nova.compute.manager [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 561.651195] env[62569]: DEBUG nova.network.neutron [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 561.687125] env[62569]: DEBUG nova.network.neutron [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.775944] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "c3f25b97-0e1e-4e78-94f0-e8087d1b55c1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.776421] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "c3f25b97-0e1e-4e78-94f0-e8087d1b55c1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.992404] env[62569]: DEBUG oslo_vmware.api [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Task: {'id': task-1249838, 'name': PowerOnVM_Task, 'duration_secs': 0.475078} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 561.993539] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 561.993832] env[62569]: DEBUG nova.compute.manager [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 561.995224] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18eab13c-0959-42e1-a49a-72bda66d787d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.999617] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a318ff46-4a42-42aa-b5f8-2cb47c9c1299 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.008475] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba2b8217-5bc8-4b25-b944-7d6e68dbc63a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.048265] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b5cd4c0-6a55-4197-9ee5-e7634147e38f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.058868] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a46f3ff6-997c-409b-81e8-018cb9089bc1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.076218] env[62569]: DEBUG nova.compute.provider_tree [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 562.191273] env[62569]: DEBUG nova.network.neutron [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.530019] env[62569]: DEBUG oslo_concurrency.lockutils [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.582846] env[62569]: DEBUG nova.scheduler.client.report [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 562.697072] env[62569]: INFO nova.compute.manager [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: e8b8b029-b20a-4af3-a5dc-1b032ce46f8d] Took 1.04 seconds to deallocate network for instance. [ 563.089902] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.667s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.090356] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 563.095462] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.914s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.600634] env[62569]: DEBUG nova.compute.utils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 563.603271] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 563.606019] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 563.681598] env[62569]: DEBUG nova.policy [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '599a3e09a44d4b6a980e34ba73f9d6e5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff4d4d8e74664c3a9fb2ab8ca80ba126', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 563.747289] env[62569]: INFO nova.scheduler.client.report [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Deleted allocations for instance e8b8b029-b20a-4af3-a5dc-1b032ce46f8d [ 564.102600] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Successfully created port: 8d126f2e-aa3d-4d1a-992e-287dcf02c111 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 564.105541] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 564.118236] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-498fca32-dafe-42c4-87b6-0a34cf69573d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.127173] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ab26665-789a-413d-9df6-60578b1a3fde {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.161235] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab33ec9-9b2c-429a-8c94-e0219c8cf25e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.172873] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524d6aeb-9da8-4327-b83e-cf35b60e25cd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.188443] env[62569]: DEBUG nova.compute.provider_tree [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 564.261962] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46ccb1bd-a2d1-4e56-96da-9c993cbb2001 tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Lock "e8b8b029-b20a-4af3-a5dc-1b032ce46f8d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.230s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.405629] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Acquiring lock "6e35fe1c-05bc-4dcc-b7ec-e9998a647739" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.405901] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Lock "6e35fe1c-05bc-4dcc-b7ec-e9998a647739" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.406118] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Acquiring lock "6e35fe1c-05bc-4dcc-b7ec-e9998a647739-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.406297] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Lock "6e35fe1c-05bc-4dcc-b7ec-e9998a647739-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.406479] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Lock "6e35fe1c-05bc-4dcc-b7ec-e9998a647739-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.409215] env[62569]: INFO nova.compute.manager [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Terminating instance [ 564.693543] env[62569]: DEBUG nova.scheduler.client.report [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 564.766908] env[62569]: DEBUG nova.compute.manager [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 564.913226] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Acquiring lock "refresh_cache-6e35fe1c-05bc-4dcc-b7ec-e9998a647739" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 564.913926] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Acquired lock "refresh_cache-6e35fe1c-05bc-4dcc-b7ec-e9998a647739" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 564.914850] env[62569]: DEBUG nova.network.neutron [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 565.103070] env[62569]: ERROR nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8d126f2e-aa3d-4d1a-992e-287dcf02c111, please check neutron logs for more information. [ 565.103070] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 565.103070] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 565.103070] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 565.103070] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 565.103070] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 565.103070] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 565.103070] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 565.103070] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.103070] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 565.103070] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.103070] env[62569]: ERROR nova.compute.manager raise self.value [ 565.103070] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 565.103070] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 565.103070] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.103070] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 565.103707] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.103707] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 565.103707] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8d126f2e-aa3d-4d1a-992e-287dcf02c111, please check neutron logs for more information. [ 565.103707] env[62569]: ERROR nova.compute.manager [ 565.103707] env[62569]: Traceback (most recent call last): [ 565.103707] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 565.103707] env[62569]: listener.cb(fileno) [ 565.103707] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.103707] env[62569]: result = function(*args, **kwargs) [ 565.103707] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 565.103707] env[62569]: return func(*args, **kwargs) [ 565.103707] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 565.103707] env[62569]: raise e [ 565.103707] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 565.103707] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 565.103707] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 565.103707] env[62569]: created_port_ids = self._update_ports_for_instance( [ 565.103707] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 565.103707] env[62569]: with excutils.save_and_reraise_exception(): [ 565.103707] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.103707] env[62569]: self.force_reraise() [ 565.103707] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.103707] env[62569]: raise self.value [ 565.103707] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 565.103707] env[62569]: updated_port = self._update_port( [ 565.103707] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.103707] env[62569]: _ensure_no_port_binding_failure(port) [ 565.103707] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.103707] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 565.104690] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 8d126f2e-aa3d-4d1a-992e-287dcf02c111, please check neutron logs for more information. [ 565.104690] env[62569]: Removing descriptor: 14 [ 565.118890] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 565.161701] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 565.162069] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 565.162479] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 565.162794] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 565.163199] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 565.163846] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 565.164105] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 565.166024] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 565.166024] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 565.166024] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 565.166024] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 565.166024] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e2982d-7f25-47e9-9aaf-ee7ec592aaf6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.174881] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-407c8eaa-fafa-4d38-b23d-7dab7a144df1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.189178] env[62569]: ERROR nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8d126f2e-aa3d-4d1a-992e-287dcf02c111, please check neutron logs for more information. [ 565.189178] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Traceback (most recent call last): [ 565.189178] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 565.189178] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] yield resources [ 565.189178] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 565.189178] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] self.driver.spawn(context, instance, image_meta, [ 565.189178] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 565.189178] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 565.189178] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 565.189178] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] vm_ref = self.build_virtual_machine(instance, [ 565.189178] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 565.189608] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] vif_infos = vmwarevif.get_vif_info(self._session, [ 565.189608] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 565.189608] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] for vif in network_info: [ 565.189608] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 565.189608] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] return self._sync_wrapper(fn, *args, **kwargs) [ 565.189608] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 565.189608] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] self.wait() [ 565.189608] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 565.189608] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] self[:] = self._gt.wait() [ 565.189608] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 565.189608] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] return self._exit_event.wait() [ 565.189608] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 565.189608] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] current.throw(*self._exc) [ 565.190073] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.190073] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] result = function(*args, **kwargs) [ 565.190073] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 565.190073] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] return func(*args, **kwargs) [ 565.190073] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 565.190073] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] raise e [ 565.190073] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 565.190073] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] nwinfo = self.network_api.allocate_for_instance( [ 565.190073] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 565.190073] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] created_port_ids = self._update_ports_for_instance( [ 565.190073] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 565.190073] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] with excutils.save_and_reraise_exception(): [ 565.190073] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.190543] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] self.force_reraise() [ 565.190543] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.190543] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] raise self.value [ 565.190543] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 565.190543] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] updated_port = self._update_port( [ 565.190543] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.190543] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] _ensure_no_port_binding_failure(port) [ 565.190543] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.190543] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] raise exception.PortBindingFailed(port_id=port['id']) [ 565.190543] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] nova.exception.PortBindingFailed: Binding failed for port 8d126f2e-aa3d-4d1a-992e-287dcf02c111, please check neutron logs for more information. [ 565.190543] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] [ 565.190543] env[62569]: INFO nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Terminating instance [ 565.198303] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.104s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.199764] env[62569]: ERROR nova.compute.manager [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 28017eb1-a7ac-4cfb-a089-ad527f6ed915, please check neutron logs for more information. [ 565.199764] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Traceback (most recent call last): [ 565.199764] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 565.199764] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] self.driver.spawn(context, instance, image_meta, [ 565.199764] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 565.199764] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] self._vmops.spawn(context, instance, image_meta, injected_files, [ 565.199764] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 565.199764] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] vm_ref = self.build_virtual_machine(instance, [ 565.199764] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 565.199764] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] vif_infos = vmwarevif.get_vif_info(self._session, [ 565.199764] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 565.200422] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] for vif in network_info: [ 565.200422] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 565.200422] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] return self._sync_wrapper(fn, *args, **kwargs) [ 565.200422] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 565.200422] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] self.wait() [ 565.200422] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 565.200422] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] self[:] = self._gt.wait() [ 565.200422] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 565.200422] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] return self._exit_event.wait() [ 565.200422] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 565.200422] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] result = hub.switch() [ 565.200422] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 565.200422] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] return self.greenlet.switch() [ 565.200838] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.200838] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] result = function(*args, **kwargs) [ 565.200838] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 565.200838] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] return func(*args, **kwargs) [ 565.200838] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 565.200838] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] raise e [ 565.200838] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 565.200838] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] nwinfo = self.network_api.allocate_for_instance( [ 565.200838] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 565.200838] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] created_port_ids = self._update_ports_for_instance( [ 565.200838] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 565.200838] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] with excutils.save_and_reraise_exception(): [ 565.200838] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.202683] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] self.force_reraise() [ 565.202683] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.202683] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] raise self.value [ 565.202683] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 565.202683] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] updated_port = self._update_port( [ 565.202683] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.202683] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] _ensure_no_port_binding_failure(port) [ 565.202683] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.202683] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] raise exception.PortBindingFailed(port_id=port['id']) [ 565.202683] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] nova.exception.PortBindingFailed: Binding failed for port 28017eb1-a7ac-4cfb-a089-ad527f6ed915, please check neutron logs for more information. [ 565.202683] env[62569]: ERROR nova.compute.manager [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] [ 565.203031] env[62569]: DEBUG nova.compute.utils [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Binding failed for port 28017eb1-a7ac-4cfb-a089-ad527f6ed915, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 565.203031] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.473s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.203031] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.203701] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.666s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.205234] env[62569]: INFO nova.compute.claims [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 565.208233] env[62569]: DEBUG nova.compute.manager [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Build of instance ba23eea7-422c-4e5d-b95e-f95589e41550 was re-scheduled: Binding failed for port 28017eb1-a7ac-4cfb-a089-ad527f6ed915, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 565.208743] env[62569]: DEBUG nova.compute.manager [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 565.209058] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Acquiring lock "refresh_cache-ba23eea7-422c-4e5d-b95e-f95589e41550" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.209058] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Acquired lock "refresh_cache-ba23eea7-422c-4e5d-b95e-f95589e41550" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.209185] env[62569]: DEBUG nova.network.neutron [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 565.246256] env[62569]: INFO nova.scheduler.client.report [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Deleted allocations for instance 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007 [ 565.290072] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.440292] env[62569]: DEBUG nova.network.neutron [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 565.491220] env[62569]: DEBUG nova.network.neutron [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.694183] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquiring lock "refresh_cache-11c53593-cf14-4e46-b546-c90a3184028e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.694467] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquired lock "refresh_cache-11c53593-cf14-4e46-b546-c90a3184028e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.694543] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 565.731164] env[62569]: DEBUG nova.network.neutron [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 565.758249] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9746e2d7-f1c0-4213-abff-1bba4f485a8e tempest-ServerDiagnosticsV248Test-1071965150 tempest-ServerDiagnosticsV248Test-1071965150-project-member] Lock "17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.532s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.849055] env[62569]: DEBUG nova.network.neutron [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.886774] env[62569]: DEBUG oslo_concurrency.lockutils [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Acquiring lock "d3274d59-6cf7-4e86-90f9-ffea49ad5342" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.887322] env[62569]: DEBUG oslo_concurrency.lockutils [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Lock "d3274d59-6cf7-4e86-90f9-ffea49ad5342" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.997532] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Releasing lock "refresh_cache-6e35fe1c-05bc-4dcc-b7ec-e9998a647739" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.997532] env[62569]: DEBUG nova.compute.manager [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 565.997532] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 565.997532] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c61c1c3-b43e-4747-8bec-b1c286b606d1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.006800] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 566.007065] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-788af5fc-d90b-42f0-ba2b-d4d0eb217774 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.015568] env[62569]: DEBUG oslo_vmware.api [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Waiting for the task: (returnval){ [ 566.015568] env[62569]: value = "task-1249839" [ 566.015568] env[62569]: _type = "Task" [ 566.015568] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.027279] env[62569]: DEBUG oslo_vmware.api [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249839, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.216090] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 566.310409] env[62569]: DEBUG nova.compute.manager [req-c526cbab-e1a6-422d-a6f9-cc446af404a1 req-5c347ea9-91b9-4b1b-980c-11592d1ebba7 service nova] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Received event network-changed-8d126f2e-aa3d-4d1a-992e-287dcf02c111 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 566.310409] env[62569]: DEBUG nova.compute.manager [req-c526cbab-e1a6-422d-a6f9-cc446af404a1 req-5c347ea9-91b9-4b1b-980c-11592d1ebba7 service nova] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Refreshing instance network info cache due to event network-changed-8d126f2e-aa3d-4d1a-992e-287dcf02c111. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 566.310409] env[62569]: DEBUG oslo_concurrency.lockutils [req-c526cbab-e1a6-422d-a6f9-cc446af404a1 req-5c347ea9-91b9-4b1b-980c-11592d1ebba7 service nova] Acquiring lock "refresh_cache-11c53593-cf14-4e46-b546-c90a3184028e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.316723] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.354580] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Releasing lock "refresh_cache-ba23eea7-422c-4e5d-b95e-f95589e41550" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.354580] env[62569]: DEBUG nova.compute.manager [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 566.354580] env[62569]: DEBUG nova.compute.manager [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 566.354758] env[62569]: DEBUG nova.network.neutron [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 566.382100] env[62569]: DEBUG nova.network.neutron [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 566.531087] env[62569]: DEBUG oslo_vmware.api [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249839, 'name': PowerOffVM_Task, 'duration_secs': 0.127705} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.531368] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 566.531530] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 566.531778] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b4e8abd9-ac63-41e6-b5eb-53746f5c9dbc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.555944] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 566.555944] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 566.555944] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Deleting the datastore file [datastore1] 6e35fe1c-05bc-4dcc-b7ec-e9998a647739 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 566.555944] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-266933a5-783e-465d-b4ab-3daa6c18ae58 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.561069] env[62569]: DEBUG oslo_vmware.api [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Waiting for the task: (returnval){ [ 566.561069] env[62569]: value = "task-1249841" [ 566.561069] env[62569]: _type = "Task" [ 566.561069] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.571826] env[62569]: DEBUG oslo_vmware.api [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249841, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.698563] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b067f713-a7e1-42db-b7ad-761cacf1ab32 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.706259] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42549b89-aac3-454e-9d35-dbca5a9b0a0a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.737769] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-010074be-a019-4278-b9c1-5f8c2ca8483c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.746666] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-904281e6-a690-4ba7-a6bd-86c25f841275 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.759953] env[62569]: DEBUG nova.compute.provider_tree [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 566.818721] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Releasing lock "refresh_cache-11c53593-cf14-4e46-b546-c90a3184028e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.819101] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 566.819348] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 566.819704] env[62569]: DEBUG oslo_concurrency.lockutils [req-c526cbab-e1a6-422d-a6f9-cc446af404a1 req-5c347ea9-91b9-4b1b-980c-11592d1ebba7 service nova] Acquired lock "refresh_cache-11c53593-cf14-4e46-b546-c90a3184028e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.819886] env[62569]: DEBUG nova.network.neutron [req-c526cbab-e1a6-422d-a6f9-cc446af404a1 req-5c347ea9-91b9-4b1b-980c-11592d1ebba7 service nova] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Refreshing network info cache for port 8d126f2e-aa3d-4d1a-992e-287dcf02c111 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 566.820968] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3382d804-e8f5-40d7-b790-161e9783282b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.831759] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f0ce1b5-9ab3-464f-8cef-fd70870be4a7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.857174] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 11c53593-cf14-4e46-b546-c90a3184028e could not be found. [ 566.857453] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 566.857733] env[62569]: INFO nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 566.857916] env[62569]: DEBUG oslo.service.loopingcall [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 566.858195] env[62569]: DEBUG nova.compute.manager [-] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 566.858294] env[62569]: DEBUG nova.network.neutron [-] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 566.880420] env[62569]: DEBUG nova.network.neutron [-] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 566.885379] env[62569]: DEBUG nova.network.neutron [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.074984] env[62569]: DEBUG oslo_vmware.api [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Task: {'id': task-1249841, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.105921} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.075735] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 567.076367] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 567.076615] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 567.077228] env[62569]: INFO nova.compute.manager [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Took 1.08 seconds to destroy the instance on the hypervisor. [ 567.077426] env[62569]: DEBUG oslo.service.loopingcall [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 567.077677] env[62569]: DEBUG nova.compute.manager [-] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 567.078087] env[62569]: DEBUG nova.network.neutron [-] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 567.100745] env[62569]: DEBUG nova.network.neutron [-] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 567.266023] env[62569]: DEBUG nova.scheduler.client.report [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 567.346491] env[62569]: DEBUG nova.network.neutron [req-c526cbab-e1a6-422d-a6f9-cc446af404a1 req-5c347ea9-91b9-4b1b-980c-11592d1ebba7 service nova] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 567.383532] env[62569]: DEBUG nova.network.neutron [-] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.389604] env[62569]: INFO nova.compute.manager [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] [instance: ba23eea7-422c-4e5d-b95e-f95589e41550] Took 1.03 seconds to deallocate network for instance. [ 567.448066] env[62569]: DEBUG nova.network.neutron [req-c526cbab-e1a6-422d-a6f9-cc446af404a1 req-5c347ea9-91b9-4b1b-980c-11592d1ebba7 service nova] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.604436] env[62569]: DEBUG nova.network.neutron [-] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.770072] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.566s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 567.770651] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 567.773929] env[62569]: DEBUG oslo_concurrency.lockutils [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.981s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.775578] env[62569]: INFO nova.compute.claims [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 567.890492] env[62569]: INFO nova.compute.manager [-] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Took 1.03 seconds to deallocate network for instance. [ 567.892789] env[62569]: DEBUG nova.compute.claims [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 567.892897] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.953761] env[62569]: DEBUG oslo_concurrency.lockutils [req-c526cbab-e1a6-422d-a6f9-cc446af404a1 req-5c347ea9-91b9-4b1b-980c-11592d1ebba7 service nova] Releasing lock "refresh_cache-11c53593-cf14-4e46-b546-c90a3184028e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.109446] env[62569]: INFO nova.compute.manager [-] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Took 1.03 seconds to deallocate network for instance. [ 568.282174] env[62569]: DEBUG nova.compute.utils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 568.286871] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 568.287053] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 568.336401] env[62569]: DEBUG nova.policy [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '599a3e09a44d4b6a980e34ba73f9d6e5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff4d4d8e74664c3a9fb2ab8ca80ba126', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 568.425167] env[62569]: INFO nova.scheduler.client.report [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Deleted allocations for instance ba23eea7-422c-4e5d-b95e-f95589e41550 [ 568.574289] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Successfully created port: faec1c73-41a9-4fba-9d98-87f46347132f {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 568.617164] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.786794] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 568.796081] env[62569]: DEBUG nova.compute.manager [req-fee601b8-ad8b-4323-8cf7-8808b6c70ff4 req-7354679f-da19-42de-b9f8-d5199d62a688 service nova] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Received event network-vif-deleted-8d126f2e-aa3d-4d1a-992e-287dcf02c111 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 568.933958] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c08713f7-f310-4f6a-bdeb-37938d9da0e9 tempest-ServerDiagnosticsTest-1398886769 tempest-ServerDiagnosticsTest-1398886769-project-member] Lock "ba23eea7-422c-4e5d-b95e-f95589e41550" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.152s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.229907] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48f0ae81-7004-4648-9c7a-ce8bbdc708e4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.238330] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13378b1a-0bcf-45cc-b385-f894bd88a65a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.268688] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-157622e4-cc65-4ec2-9d9c-61bb893c7339 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.275789] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70330bb3-9d6e-4807-a0f2-6942b442625f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.288738] env[62569]: DEBUG nova.compute.provider_tree [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 569.440018] env[62569]: DEBUG nova.compute.manager [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 569.602766] env[62569]: ERROR nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port faec1c73-41a9-4fba-9d98-87f46347132f, please check neutron logs for more information. [ 569.602766] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 569.602766] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 569.602766] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 569.602766] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 569.602766] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 569.602766] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 569.602766] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 569.602766] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.602766] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 569.602766] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.602766] env[62569]: ERROR nova.compute.manager raise self.value [ 569.602766] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 569.602766] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 569.602766] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.602766] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 569.603364] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.603364] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 569.603364] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port faec1c73-41a9-4fba-9d98-87f46347132f, please check neutron logs for more information. [ 569.603364] env[62569]: ERROR nova.compute.manager [ 569.603364] env[62569]: Traceback (most recent call last): [ 569.603364] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 569.603364] env[62569]: listener.cb(fileno) [ 569.603364] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.603364] env[62569]: result = function(*args, **kwargs) [ 569.603364] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 569.603364] env[62569]: return func(*args, **kwargs) [ 569.603364] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 569.603364] env[62569]: raise e [ 569.603364] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 569.603364] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 569.603364] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 569.603364] env[62569]: created_port_ids = self._update_ports_for_instance( [ 569.603364] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 569.603364] env[62569]: with excutils.save_and_reraise_exception(): [ 569.603364] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.603364] env[62569]: self.force_reraise() [ 569.603364] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.603364] env[62569]: raise self.value [ 569.603364] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 569.603364] env[62569]: updated_port = self._update_port( [ 569.603364] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.603364] env[62569]: _ensure_no_port_binding_failure(port) [ 569.603364] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.603364] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 569.604267] env[62569]: nova.exception.PortBindingFailed: Binding failed for port faec1c73-41a9-4fba-9d98-87f46347132f, please check neutron logs for more information. [ 569.604267] env[62569]: Removing descriptor: 14 [ 569.794380] env[62569]: DEBUG nova.scheduler.client.report [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 569.801765] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 569.836020] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 569.836253] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 569.836410] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 569.836784] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 569.836784] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 569.836893] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 569.837160] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 569.837424] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 569.837654] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 569.837947] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 569.838241] env[62569]: DEBUG nova.virt.hardware [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 569.839592] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9345acd3-b225-4718-8dc0-0db19222cdb2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.849686] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ecbdd06-3826-4077-8491-d2bd380a02d0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.866020] env[62569]: ERROR nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port faec1c73-41a9-4fba-9d98-87f46347132f, please check neutron logs for more information. [ 569.866020] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Traceback (most recent call last): [ 569.866020] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 569.866020] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] yield resources [ 569.866020] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 569.866020] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] self.driver.spawn(context, instance, image_meta, [ 569.866020] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 569.866020] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 569.866020] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 569.866020] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] vm_ref = self.build_virtual_machine(instance, [ 569.866020] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 569.866509] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] vif_infos = vmwarevif.get_vif_info(self._session, [ 569.866509] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 569.866509] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] for vif in network_info: [ 569.866509] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 569.866509] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] return self._sync_wrapper(fn, *args, **kwargs) [ 569.866509] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 569.866509] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] self.wait() [ 569.866509] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 569.866509] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] self[:] = self._gt.wait() [ 569.866509] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 569.866509] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] return self._exit_event.wait() [ 569.866509] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 569.866509] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] current.throw(*self._exc) [ 569.866959] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.866959] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] result = function(*args, **kwargs) [ 569.866959] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 569.866959] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] return func(*args, **kwargs) [ 569.866959] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 569.866959] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] raise e [ 569.866959] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 569.866959] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] nwinfo = self.network_api.allocate_for_instance( [ 569.866959] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 569.866959] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] created_port_ids = self._update_ports_for_instance( [ 569.866959] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 569.866959] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] with excutils.save_and_reraise_exception(): [ 569.866959] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.867470] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] self.force_reraise() [ 569.867470] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.867470] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] raise self.value [ 569.867470] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 569.867470] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] updated_port = self._update_port( [ 569.867470] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.867470] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] _ensure_no_port_binding_failure(port) [ 569.867470] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.867470] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] raise exception.PortBindingFailed(port_id=port['id']) [ 569.867470] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] nova.exception.PortBindingFailed: Binding failed for port faec1c73-41a9-4fba-9d98-87f46347132f, please check neutron logs for more information. [ 569.867470] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] [ 569.867470] env[62569]: INFO nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Terminating instance [ 569.960924] env[62569]: DEBUG oslo_concurrency.lockutils [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.307125] env[62569]: DEBUG oslo_concurrency.lockutils [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.533s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.307444] env[62569]: DEBUG nova.compute.manager [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 570.311625] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.424s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.372146] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquiring lock "refresh_cache-a9224414-6d20-4d5f-be5d-2b893117f4b8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 570.372146] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquired lock "refresh_cache-a9224414-6d20-4d5f-be5d-2b893117f4b8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 570.372146] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 570.817479] env[62569]: DEBUG nova.compute.utils [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 570.827015] env[62569]: DEBUG nova.compute.manager [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 570.827015] env[62569]: DEBUG nova.network.neutron [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 570.905893] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 570.951955] env[62569]: DEBUG nova.policy [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7948a83ec06e43a284f4882fd9951484', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fd57a13c3cab4801bad22a91776c620d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 571.084850] env[62569]: DEBUG nova.compute.manager [req-eaca29ac-7838-4081-9d26-3f9bc0fe5712 req-2a98521e-2bc8-4346-8924-f8b1cb8c9dcf service nova] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Received event network-changed-faec1c73-41a9-4fba-9d98-87f46347132f {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 571.084924] env[62569]: DEBUG nova.compute.manager [req-eaca29ac-7838-4081-9d26-3f9bc0fe5712 req-2a98521e-2bc8-4346-8924-f8b1cb8c9dcf service nova] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Refreshing instance network info cache due to event network-changed-faec1c73-41a9-4fba-9d98-87f46347132f. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 571.085100] env[62569]: DEBUG oslo_concurrency.lockutils [req-eaca29ac-7838-4081-9d26-3f9bc0fe5712 req-2a98521e-2bc8-4346-8924-f8b1cb8c9dcf service nova] Acquiring lock "refresh_cache-a9224414-6d20-4d5f-be5d-2b893117f4b8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 571.097549] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.238554] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6a590bc-14da-4805-8ec3-464b6e882051 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.246211] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc9550b0-b502-4709-b09c-d072fa423054 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.276232] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac8d60fb-735c-4be9-84cf-50b64cb0e564 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.283884] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42fcc682-04c2-401d-b0d2-8e97281db5d8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.297112] env[62569]: DEBUG nova.compute.provider_tree [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 571.332063] env[62569]: DEBUG nova.compute.manager [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 571.526126] env[62569]: DEBUG nova.network.neutron [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Successfully created port: 04c1526d-781c-4556-ad2a-18d15aa064e0 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 571.600341] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Releasing lock "refresh_cache-a9224414-6d20-4d5f-be5d-2b893117f4b8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.600715] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 571.600890] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 571.602177] env[62569]: DEBUG oslo_concurrency.lockutils [req-eaca29ac-7838-4081-9d26-3f9bc0fe5712 req-2a98521e-2bc8-4346-8924-f8b1cb8c9dcf service nova] Acquired lock "refresh_cache-a9224414-6d20-4d5f-be5d-2b893117f4b8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.602177] env[62569]: DEBUG nova.network.neutron [req-eaca29ac-7838-4081-9d26-3f9bc0fe5712 req-2a98521e-2bc8-4346-8924-f8b1cb8c9dcf service nova] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Refreshing network info cache for port faec1c73-41a9-4fba-9d98-87f46347132f {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 571.602599] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-24f8efac-edcd-4f2d-985f-dd8f5d0f62a3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.617337] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a865e3-e36e-4fb9-85a3-cf97efb60218 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.645751] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a9224414-6d20-4d5f-be5d-2b893117f4b8 could not be found. [ 571.645751] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 571.645751] env[62569]: INFO nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 571.645751] env[62569]: DEBUG oslo.service.loopingcall [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 571.645751] env[62569]: DEBUG nova.compute.manager [-] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 571.645751] env[62569]: DEBUG nova.network.neutron [-] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 571.663507] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cacfbb43-9283-453e-9044-851bf0c24b47 tempest-ServersListShow296Test-1909070751 tempest-ServersListShow296Test-1909070751-project-member] Acquiring lock "d0fbb20f-bfcc-491e-a7ec-e7e04a0d11fe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.663793] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cacfbb43-9283-453e-9044-851bf0c24b47 tempest-ServersListShow296Test-1909070751 tempest-ServersListShow296Test-1909070751-project-member] Lock "d0fbb20f-bfcc-491e-a7ec-e7e04a0d11fe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.679206] env[62569]: DEBUG nova.network.neutron [-] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 571.803612] env[62569]: DEBUG nova.scheduler.client.report [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 572.137784] env[62569]: DEBUG nova.network.neutron [req-eaca29ac-7838-4081-9d26-3f9bc0fe5712 req-2a98521e-2bc8-4346-8924-f8b1cb8c9dcf service nova] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 572.182984] env[62569]: DEBUG nova.network.neutron [-] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.309169] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.995s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.309169] env[62569]: ERROR nova.compute.manager [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2a118c7a-8fdc-4948-9ec7-facf3bf84bd8, please check neutron logs for more information. [ 572.309169] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Traceback (most recent call last): [ 572.309169] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 572.309169] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] self.driver.spawn(context, instance, image_meta, [ 572.309169] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 572.309169] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 572.309169] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 572.309169] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] vm_ref = self.build_virtual_machine(instance, [ 572.310826] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 572.310826] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] vif_infos = vmwarevif.get_vif_info(self._session, [ 572.310826] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 572.310826] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] for vif in network_info: [ 572.310826] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 572.310826] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] return self._sync_wrapper(fn, *args, **kwargs) [ 572.310826] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 572.310826] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] self.wait() [ 572.310826] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 572.310826] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] self[:] = self._gt.wait() [ 572.310826] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 572.310826] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] return self._exit_event.wait() [ 572.310826] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 572.311371] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] result = hub.switch() [ 572.311371] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 572.311371] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] return self.greenlet.switch() [ 572.311371] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.311371] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] result = function(*args, **kwargs) [ 572.311371] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 572.311371] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] return func(*args, **kwargs) [ 572.311371] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 572.311371] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] raise e [ 572.311371] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 572.311371] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] nwinfo = self.network_api.allocate_for_instance( [ 572.311371] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.311371] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] created_port_ids = self._update_ports_for_instance( [ 572.311777] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.311777] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] with excutils.save_and_reraise_exception(): [ 572.311777] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.311777] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] self.force_reraise() [ 572.311777] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.311777] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] raise self.value [ 572.311777] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.311777] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] updated_port = self._update_port( [ 572.311777] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.311777] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] _ensure_no_port_binding_failure(port) [ 572.311777] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.311777] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] raise exception.PortBindingFailed(port_id=port['id']) [ 572.312315] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] nova.exception.PortBindingFailed: Binding failed for port 2a118c7a-8fdc-4948-9ec7-facf3bf84bd8, please check neutron logs for more information. [ 572.312315] env[62569]: ERROR nova.compute.manager [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] [ 572.312315] env[62569]: DEBUG nova.compute.utils [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Binding failed for port 2a118c7a-8fdc-4948-9ec7-facf3bf84bd8, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 572.312770] env[62569]: DEBUG nova.network.neutron [req-eaca29ac-7838-4081-9d26-3f9bc0fe5712 req-2a98521e-2bc8-4346-8924-f8b1cb8c9dcf service nova] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.313906] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.828s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.317532] env[62569]: INFO nova.compute.claims [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 572.322753] env[62569]: DEBUG nova.compute.manager [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Build of instance 9fa91603-f812-4d93-bef1-58e455f2c2e8 was re-scheduled: Binding failed for port 2a118c7a-8fdc-4948-9ec7-facf3bf84bd8, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 572.323530] env[62569]: DEBUG nova.compute.manager [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 572.324162] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Acquiring lock "refresh_cache-9fa91603-f812-4d93-bef1-58e455f2c2e8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.325128] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Acquired lock "refresh_cache-9fa91603-f812-4d93-bef1-58e455f2c2e8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.325128] env[62569]: DEBUG nova.network.neutron [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 572.344060] env[62569]: DEBUG nova.compute.manager [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 572.386667] env[62569]: DEBUG nova.virt.hardware [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 572.386921] env[62569]: DEBUG nova.virt.hardware [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 572.387086] env[62569]: DEBUG nova.virt.hardware [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 572.387315] env[62569]: DEBUG nova.virt.hardware [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 572.389514] env[62569]: DEBUG nova.virt.hardware [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 572.389514] env[62569]: DEBUG nova.virt.hardware [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 572.389514] env[62569]: DEBUG nova.virt.hardware [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 572.389514] env[62569]: DEBUG nova.virt.hardware [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 572.389514] env[62569]: DEBUG nova.virt.hardware [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 572.389817] env[62569]: DEBUG nova.virt.hardware [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 572.389817] env[62569]: DEBUG nova.virt.hardware [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 572.390625] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb391f96-e27d-4da0-8ba5-ca4b15d7b1ec {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.400161] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc1ad04-e37a-4900-b746-64c09339db10 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.686351] env[62569]: INFO nova.compute.manager [-] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Took 1.04 seconds to deallocate network for instance. [ 572.694621] env[62569]: DEBUG nova.compute.claims [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 572.694815] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.827333] env[62569]: DEBUG oslo_concurrency.lockutils [req-eaca29ac-7838-4081-9d26-3f9bc0fe5712 req-2a98521e-2bc8-4346-8924-f8b1cb8c9dcf service nova] Releasing lock "refresh_cache-a9224414-6d20-4d5f-be5d-2b893117f4b8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 572.829056] env[62569]: DEBUG nova.compute.manager [req-eaca29ac-7838-4081-9d26-3f9bc0fe5712 req-2a98521e-2bc8-4346-8924-f8b1cb8c9dcf service nova] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Received event network-vif-deleted-faec1c73-41a9-4fba-9d98-87f46347132f {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 572.876577] env[62569]: DEBUG nova.network.neutron [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.050619] env[62569]: DEBUG nova.network.neutron [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.476146] env[62569]: ERROR nova.compute.manager [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 04c1526d-781c-4556-ad2a-18d15aa064e0, please check neutron logs for more information. [ 573.476146] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 573.476146] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 573.476146] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 573.476146] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 573.476146] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 573.476146] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 573.476146] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 573.476146] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.476146] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 573.476146] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.476146] env[62569]: ERROR nova.compute.manager raise self.value [ 573.476146] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 573.476146] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 573.476146] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.476146] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 573.476924] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.476924] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 573.476924] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 04c1526d-781c-4556-ad2a-18d15aa064e0, please check neutron logs for more information. [ 573.476924] env[62569]: ERROR nova.compute.manager [ 573.476924] env[62569]: Traceback (most recent call last): [ 573.476924] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 573.476924] env[62569]: listener.cb(fileno) [ 573.476924] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.476924] env[62569]: result = function(*args, **kwargs) [ 573.476924] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.476924] env[62569]: return func(*args, **kwargs) [ 573.476924] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 573.476924] env[62569]: raise e [ 573.476924] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 573.476924] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 573.476924] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 573.476924] env[62569]: created_port_ids = self._update_ports_for_instance( [ 573.476924] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 573.476924] env[62569]: with excutils.save_and_reraise_exception(): [ 573.476924] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.476924] env[62569]: self.force_reraise() [ 573.476924] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.476924] env[62569]: raise self.value [ 573.476924] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 573.476924] env[62569]: updated_port = self._update_port( [ 573.476924] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.476924] env[62569]: _ensure_no_port_binding_failure(port) [ 573.476924] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.476924] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 573.477922] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 04c1526d-781c-4556-ad2a-18d15aa064e0, please check neutron logs for more information. [ 573.477922] env[62569]: Removing descriptor: 14 [ 573.477922] env[62569]: ERROR nova.compute.manager [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 04c1526d-781c-4556-ad2a-18d15aa064e0, please check neutron logs for more information. [ 573.477922] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Traceback (most recent call last): [ 573.477922] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 573.477922] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] yield resources [ 573.477922] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 573.477922] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] self.driver.spawn(context, instance, image_meta, [ 573.477922] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 573.477922] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 573.477922] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 573.477922] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] vm_ref = self.build_virtual_machine(instance, [ 573.478364] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 573.478364] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] vif_infos = vmwarevif.get_vif_info(self._session, [ 573.478364] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 573.478364] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] for vif in network_info: [ 573.478364] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 573.478364] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] return self._sync_wrapper(fn, *args, **kwargs) [ 573.478364] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 573.478364] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] self.wait() [ 573.478364] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 573.478364] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] self[:] = self._gt.wait() [ 573.478364] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 573.478364] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] return self._exit_event.wait() [ 573.478364] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 573.479723] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] result = hub.switch() [ 573.479723] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 573.479723] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] return self.greenlet.switch() [ 573.479723] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.479723] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] result = function(*args, **kwargs) [ 573.479723] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.479723] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] return func(*args, **kwargs) [ 573.479723] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 573.479723] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] raise e [ 573.479723] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 573.479723] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] nwinfo = self.network_api.allocate_for_instance( [ 573.479723] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 573.479723] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] created_port_ids = self._update_ports_for_instance( [ 573.480318] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 573.480318] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] with excutils.save_and_reraise_exception(): [ 573.480318] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.480318] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] self.force_reraise() [ 573.480318] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.480318] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] raise self.value [ 573.480318] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 573.480318] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] updated_port = self._update_port( [ 573.480318] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.480318] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] _ensure_no_port_binding_failure(port) [ 573.480318] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.480318] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] raise exception.PortBindingFailed(port_id=port['id']) [ 573.480790] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] nova.exception.PortBindingFailed: Binding failed for port 04c1526d-781c-4556-ad2a-18d15aa064e0, please check neutron logs for more information. [ 573.480790] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] [ 573.480790] env[62569]: INFO nova.compute.manager [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Terminating instance [ 573.553220] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Releasing lock "refresh_cache-9fa91603-f812-4d93-bef1-58e455f2c2e8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 573.554505] env[62569]: DEBUG nova.compute.manager [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 573.554505] env[62569]: DEBUG nova.compute.manager [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 573.554505] env[62569]: DEBUG nova.network.neutron [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 573.576885] env[62569]: DEBUG nova.network.neutron [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.604174] env[62569]: DEBUG nova.compute.manager [req-4013a421-2338-4e55-beb8-ce9c1379592e req-053dba6b-8ea4-4c70-aab4-78ee057179b8 service nova] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Received event network-changed-04c1526d-781c-4556-ad2a-18d15aa064e0 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 573.604575] env[62569]: DEBUG nova.compute.manager [req-4013a421-2338-4e55-beb8-ce9c1379592e req-053dba6b-8ea4-4c70-aab4-78ee057179b8 service nova] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Refreshing instance network info cache due to event network-changed-04c1526d-781c-4556-ad2a-18d15aa064e0. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 573.604818] env[62569]: DEBUG oslo_concurrency.lockutils [req-4013a421-2338-4e55-beb8-ce9c1379592e req-053dba6b-8ea4-4c70-aab4-78ee057179b8 service nova] Acquiring lock "refresh_cache-49fdf371-5278-442b-b8cc-9a0fc22947b4" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.605015] env[62569]: DEBUG oslo_concurrency.lockutils [req-4013a421-2338-4e55-beb8-ce9c1379592e req-053dba6b-8ea4-4c70-aab4-78ee057179b8 service nova] Acquired lock "refresh_cache-49fdf371-5278-442b-b8cc-9a0fc22947b4" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.605398] env[62569]: DEBUG nova.network.neutron [req-4013a421-2338-4e55-beb8-ce9c1379592e req-053dba6b-8ea4-4c70-aab4-78ee057179b8 service nova] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Refreshing network info cache for port 04c1526d-781c-4556-ad2a-18d15aa064e0 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 573.756545] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af580819-41d7-43f4-be68-db695d34f127 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.764521] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2abd6c6-1652-49cb-b477-3e7cd18d4f9d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.797063] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d11b745-2654-48b3-817e-d1a5569bffec {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.804665] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a75ebfa-ef2e-4295-80a2-f91b2f12e68f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.819884] env[62569]: DEBUG nova.compute.provider_tree [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 573.982372] env[62569]: DEBUG oslo_concurrency.lockutils [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Acquiring lock "refresh_cache-49fdf371-5278-442b-b8cc-9a0fc22947b4" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.079030] env[62569]: DEBUG nova.network.neutron [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.134406] env[62569]: DEBUG nova.network.neutron [req-4013a421-2338-4e55-beb8-ce9c1379592e req-053dba6b-8ea4-4c70-aab4-78ee057179b8 service nova] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.280028] env[62569]: DEBUG nova.network.neutron [req-4013a421-2338-4e55-beb8-ce9c1379592e req-053dba6b-8ea4-4c70-aab4-78ee057179b8 service nova] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.322673] env[62569]: DEBUG nova.scheduler.client.report [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 574.584514] env[62569]: INFO nova.compute.manager [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] [instance: 9fa91603-f812-4d93-bef1-58e455f2c2e8] Took 1.03 seconds to deallocate network for instance. [ 574.784645] env[62569]: DEBUG oslo_concurrency.lockutils [req-4013a421-2338-4e55-beb8-ce9c1379592e req-053dba6b-8ea4-4c70-aab4-78ee057179b8 service nova] Releasing lock "refresh_cache-49fdf371-5278-442b-b8cc-9a0fc22947b4" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.784645] env[62569]: DEBUG oslo_concurrency.lockutils [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Acquired lock "refresh_cache-49fdf371-5278-442b-b8cc-9a0fc22947b4" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.784645] env[62569]: DEBUG nova.network.neutron [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 574.835671] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.518s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.836644] env[62569]: DEBUG nova.compute.manager [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 574.841278] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.534s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.319295] env[62569]: DEBUG nova.network.neutron [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 575.349539] env[62569]: DEBUG nova.compute.utils [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 575.351559] env[62569]: DEBUG nova.compute.manager [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 575.352749] env[62569]: DEBUG nova.network.neutron [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 575.456060] env[62569]: DEBUG nova.policy [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6afea4ca95ea458aa26ad4ef74921072', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '73cf75e06b424da98d1bad15070ea90b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 575.458099] env[62569]: DEBUG nova.network.neutron [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.617934] env[62569]: INFO nova.scheduler.client.report [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Deleted allocations for instance 9fa91603-f812-4d93-bef1-58e455f2c2e8 [ 575.817353] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1739179-b740-44e9-b265-5172e4da0cdc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.827103] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9718739a-6cb9-4ae7-9cad-861a194fb496 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.867192] env[62569]: DEBUG nova.compute.manager [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 575.871156] env[62569]: DEBUG nova.network.neutron [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Successfully created port: 7f65cf1d-d031-4fe4-bc72-31049bfa7240 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 575.877485] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af6948a-2ce4-42a4-b4f1-75ab8cfbbd2d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.887864] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec08f0f0-d61c-405a-8173-51a9bc0a4eaa {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.902615] env[62569]: DEBUG nova.compute.provider_tree [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 575.962370] env[62569]: DEBUG oslo_concurrency.lockutils [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Releasing lock "refresh_cache-49fdf371-5278-442b-b8cc-9a0fc22947b4" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.962817] env[62569]: DEBUG nova.compute.manager [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 575.963020] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 575.963309] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b0834502-5131-4615-ab02-59b51aac2198 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.972990] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b0df944-eba9-4c78-9871-5e4570e3e7a7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.997071] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 49fdf371-5278-442b-b8cc-9a0fc22947b4 could not be found. [ 575.997272] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 575.998170] env[62569]: INFO nova.compute.manager [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Took 0.03 seconds to destroy the instance on the hypervisor. [ 575.998170] env[62569]: DEBUG oslo.service.loopingcall [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 575.998170] env[62569]: DEBUG nova.compute.manager [-] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 575.998170] env[62569]: DEBUG nova.network.neutron [-] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 576.020476] env[62569]: DEBUG nova.network.neutron [-] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 576.143547] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab595f99-b44e-4342-8a88-3be0dcffc30d tempest-ServersAdminTestJSON-2108863720 tempest-ServersAdminTestJSON-2108863720-project-member] Lock "9fa91603-f812-4d93-bef1-58e455f2c2e8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.250s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.386247] env[62569]: DEBUG nova.compute.manager [req-6a23b0b7-5b79-40fd-b0b0-e750d17024b1 req-26623d67-167d-4180-b5a2-6382b9cb8a77 service nova] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Received event network-vif-deleted-04c1526d-781c-4556-ad2a-18d15aa064e0 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 576.407930] env[62569]: DEBUG nova.scheduler.client.report [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 576.522765] env[62569]: DEBUG nova.network.neutron [-] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.553307] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Acquiring lock "4f45ccc2-0352-4c25-9a92-aee8f051c0b5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.553581] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Lock "4f45ccc2-0352-4c25-9a92-aee8f051c0b5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.646534] env[62569]: DEBUG nova.compute.manager [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 576.887983] env[62569]: DEBUG nova.compute.manager [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 576.915495] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.076s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.916436] env[62569]: ERROR nova.compute.manager [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6ed7b301-12d9-4c35-880f-58425e991502, please check neutron logs for more information. [ 576.916436] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Traceback (most recent call last): [ 576.916436] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 576.916436] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] self.driver.spawn(context, instance, image_meta, [ 576.916436] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 576.916436] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 576.916436] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 576.916436] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] vm_ref = self.build_virtual_machine(instance, [ 576.916436] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 576.916436] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] vif_infos = vmwarevif.get_vif_info(self._session, [ 576.916436] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 576.919464] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] for vif in network_info: [ 576.919464] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 576.919464] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] return self._sync_wrapper(fn, *args, **kwargs) [ 576.919464] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 576.919464] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] self.wait() [ 576.919464] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 576.919464] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] self[:] = self._gt.wait() [ 576.919464] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 576.919464] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] return self._exit_event.wait() [ 576.919464] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 576.919464] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] result = hub.switch() [ 576.919464] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 576.919464] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] return self.greenlet.switch() [ 576.919908] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 576.919908] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] result = function(*args, **kwargs) [ 576.919908] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 576.919908] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] return func(*args, **kwargs) [ 576.919908] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 576.919908] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] raise e [ 576.919908] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 576.919908] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] nwinfo = self.network_api.allocate_for_instance( [ 576.919908] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 576.919908] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] created_port_ids = self._update_ports_for_instance( [ 576.919908] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 576.919908] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] with excutils.save_and_reraise_exception(): [ 576.919908] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.920772] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] self.force_reraise() [ 576.920772] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.920772] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] raise self.value [ 576.920772] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 576.920772] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] updated_port = self._update_port( [ 576.920772] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.920772] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] _ensure_no_port_binding_failure(port) [ 576.920772] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.920772] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] raise exception.PortBindingFailed(port_id=port['id']) [ 576.920772] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] nova.exception.PortBindingFailed: Binding failed for port 6ed7b301-12d9-4c35-880f-58425e991502, please check neutron logs for more information. [ 576.920772] env[62569]: ERROR nova.compute.manager [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] [ 576.921315] env[62569]: DEBUG nova.compute.utils [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Binding failed for port 6ed7b301-12d9-4c35-880f-58425e991502, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 576.922833] env[62569]: DEBUG oslo_concurrency.lockutils [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.562s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.925375] env[62569]: INFO nova.compute.claims [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 576.931572] env[62569]: DEBUG nova.virt.hardware [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 576.931812] env[62569]: DEBUG nova.virt.hardware [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 576.931962] env[62569]: DEBUG nova.virt.hardware [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 576.934146] env[62569]: DEBUG nova.virt.hardware [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 576.934146] env[62569]: DEBUG nova.virt.hardware [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 576.934300] env[62569]: DEBUG nova.virt.hardware [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 576.934543] env[62569]: DEBUG nova.virt.hardware [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 576.934706] env[62569]: DEBUG nova.virt.hardware [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 576.934872] env[62569]: DEBUG nova.virt.hardware [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 576.935044] env[62569]: DEBUG nova.virt.hardware [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 576.935217] env[62569]: DEBUG nova.virt.hardware [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 576.935783] env[62569]: DEBUG nova.compute.manager [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Build of instance 5b1117bd-6210-4df2-b828-f76951752f2f was re-scheduled: Binding failed for port 6ed7b301-12d9-4c35-880f-58425e991502, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 576.936284] env[62569]: DEBUG nova.compute.manager [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 576.936532] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Acquiring lock "refresh_cache-5b1117bd-6210-4df2-b828-f76951752f2f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.937741] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Acquired lock "refresh_cache-5b1117bd-6210-4df2-b828-f76951752f2f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.938091] env[62569]: DEBUG nova.network.neutron [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 576.942022] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7dde87f-70f6-4961-a48b-3bdd63fecc07 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.950689] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c6d97c7-daf6-4a17-9b84-0fbd0fab165d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.026500] env[62569]: INFO nova.compute.manager [-] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Took 1.03 seconds to deallocate network for instance. [ 577.031218] env[62569]: DEBUG nova.compute.claims [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 577.031551] env[62569]: DEBUG oslo_concurrency.lockutils [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.188778] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.504077] env[62569]: DEBUG nova.network.neutron [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.757893] env[62569]: DEBUG nova.network.neutron [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.031256] env[62569]: DEBUG oslo_concurrency.lockutils [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Acquiring lock "bda4e4c4-b1e2-4441-8aa5-6d1b5975647c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.031256] env[62569]: DEBUG oslo_concurrency.lockutils [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Lock "bda4e4c4-b1e2-4441-8aa5-6d1b5975647c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.120297] env[62569]: ERROR nova.compute.manager [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7f65cf1d-d031-4fe4-bc72-31049bfa7240, please check neutron logs for more information. [ 578.120297] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 578.120297] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 578.120297] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 578.120297] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 578.120297] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 578.120297] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 578.120297] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 578.120297] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.120297] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 578.120297] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.120297] env[62569]: ERROR nova.compute.manager raise self.value [ 578.120297] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 578.120297] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 578.120297] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.120297] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 578.120919] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.120919] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 578.120919] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7f65cf1d-d031-4fe4-bc72-31049bfa7240, please check neutron logs for more information. [ 578.120919] env[62569]: ERROR nova.compute.manager [ 578.120919] env[62569]: Traceback (most recent call last): [ 578.120919] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 578.120919] env[62569]: listener.cb(fileno) [ 578.120919] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 578.120919] env[62569]: result = function(*args, **kwargs) [ 578.120919] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 578.120919] env[62569]: return func(*args, **kwargs) [ 578.120919] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 578.120919] env[62569]: raise e [ 578.120919] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 578.120919] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 578.120919] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 578.120919] env[62569]: created_port_ids = self._update_ports_for_instance( [ 578.120919] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 578.120919] env[62569]: with excutils.save_and_reraise_exception(): [ 578.120919] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.120919] env[62569]: self.force_reraise() [ 578.120919] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.120919] env[62569]: raise self.value [ 578.120919] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 578.120919] env[62569]: updated_port = self._update_port( [ 578.120919] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.120919] env[62569]: _ensure_no_port_binding_failure(port) [ 578.120919] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.120919] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 578.121923] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 7f65cf1d-d031-4fe4-bc72-31049bfa7240, please check neutron logs for more information. [ 578.121923] env[62569]: Removing descriptor: 17 [ 578.121923] env[62569]: ERROR nova.compute.manager [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7f65cf1d-d031-4fe4-bc72-31049bfa7240, please check neutron logs for more information. [ 578.121923] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Traceback (most recent call last): [ 578.121923] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 578.121923] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] yield resources [ 578.121923] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 578.121923] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] self.driver.spawn(context, instance, image_meta, [ 578.121923] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 578.121923] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] self._vmops.spawn(context, instance, image_meta, injected_files, [ 578.121923] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 578.121923] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] vm_ref = self.build_virtual_machine(instance, [ 578.122406] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 578.122406] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] vif_infos = vmwarevif.get_vif_info(self._session, [ 578.122406] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 578.122406] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] for vif in network_info: [ 578.122406] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 578.122406] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] return self._sync_wrapper(fn, *args, **kwargs) [ 578.122406] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 578.122406] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] self.wait() [ 578.122406] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 578.122406] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] self[:] = self._gt.wait() [ 578.122406] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 578.122406] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] return self._exit_event.wait() [ 578.122406] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 578.122879] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] result = hub.switch() [ 578.122879] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 578.122879] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] return self.greenlet.switch() [ 578.122879] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 578.122879] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] result = function(*args, **kwargs) [ 578.122879] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 578.122879] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] return func(*args, **kwargs) [ 578.122879] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 578.122879] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] raise e [ 578.122879] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 578.122879] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] nwinfo = self.network_api.allocate_for_instance( [ 578.122879] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 578.122879] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] created_port_ids = self._update_ports_for_instance( [ 578.123356] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 578.123356] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] with excutils.save_and_reraise_exception(): [ 578.123356] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.123356] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] self.force_reraise() [ 578.123356] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.123356] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] raise self.value [ 578.123356] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 578.123356] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] updated_port = self._update_port( [ 578.123356] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.123356] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] _ensure_no_port_binding_failure(port) [ 578.123356] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.123356] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] raise exception.PortBindingFailed(port_id=port['id']) [ 578.124429] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] nova.exception.PortBindingFailed: Binding failed for port 7f65cf1d-d031-4fe4-bc72-31049bfa7240, please check neutron logs for more information. [ 578.124429] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] [ 578.124429] env[62569]: INFO nova.compute.manager [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Terminating instance [ 578.260012] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Releasing lock "refresh_cache-5b1117bd-6210-4df2-b828-f76951752f2f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.260304] env[62569]: DEBUG nova.compute.manager [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 578.260518] env[62569]: DEBUG nova.compute.manager [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 578.260694] env[62569]: DEBUG nova.network.neutron [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 578.284355] env[62569]: DEBUG nova.network.neutron [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 578.440467] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c5ed5e-2a8d-45ff-bf9b-59f67ce32aac {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.449977] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1165b077-9a2f-4873-9ec2-0b3dbce0c593 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.487380] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b400978a-4599-4c89-98b3-d326d1e5a8d4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.495409] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74132a2f-14f5-44be-b278-8b9b33578451 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.509487] env[62569]: DEBUG nova.compute.provider_tree [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 578.626670] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Acquiring lock "refresh_cache-ef52ac4a-fa98-4af8-b08a-b34cae526850" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.626748] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Acquired lock "refresh_cache-ef52ac4a-fa98-4af8-b08a-b34cae526850" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.628123] env[62569]: DEBUG nova.network.neutron [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 578.788766] env[62569]: DEBUG nova.network.neutron [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.012220] env[62569]: DEBUG nova.scheduler.client.report [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 579.172596] env[62569]: DEBUG nova.network.neutron [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 579.295853] env[62569]: INFO nova.compute.manager [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 5b1117bd-6210-4df2-b828-f76951752f2f] Took 1.03 seconds to deallocate network for instance. [ 579.303475] env[62569]: DEBUG nova.network.neutron [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.336843] env[62569]: DEBUG nova.compute.manager [req-ea84cb7f-5d7a-40f7-8cc1-95d94a039b23 req-9f392b4c-44fe-4f24-b6d3-84ae9dbef6f2 service nova] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Received event network-changed-7f65cf1d-d031-4fe4-bc72-31049bfa7240 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 579.337053] env[62569]: DEBUG nova.compute.manager [req-ea84cb7f-5d7a-40f7-8cc1-95d94a039b23 req-9f392b4c-44fe-4f24-b6d3-84ae9dbef6f2 service nova] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Refreshing instance network info cache due to event network-changed-7f65cf1d-d031-4fe4-bc72-31049bfa7240. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 579.337244] env[62569]: DEBUG oslo_concurrency.lockutils [req-ea84cb7f-5d7a-40f7-8cc1-95d94a039b23 req-9f392b4c-44fe-4f24-b6d3-84ae9dbef6f2 service nova] Acquiring lock "refresh_cache-ef52ac4a-fa98-4af8-b08a-b34cae526850" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.354689] env[62569]: DEBUG oslo_concurrency.lockutils [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Acquiring lock "a930be54-d8c9-4407-9d49-f067defc65e9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.354938] env[62569]: DEBUG oslo_concurrency.lockutils [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Lock "a930be54-d8c9-4407-9d49-f067defc65e9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.424103] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Acquiring lock "34173a5f-8e4d-4646-b3e3-d537ffe752bb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.425184] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Lock "34173a5f-8e4d-4646-b3e3-d537ffe752bb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.517612] env[62569]: DEBUG oslo_concurrency.lockutils [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.595s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.518153] env[62569]: DEBUG nova.compute.manager [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 579.521446] env[62569]: DEBUG oslo_concurrency.lockutils [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 26.892s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.521637] env[62569]: DEBUG nova.objects.instance [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62569) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 579.806033] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Releasing lock "refresh_cache-ef52ac4a-fa98-4af8-b08a-b34cae526850" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.808781] env[62569]: DEBUG nova.compute.manager [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 579.808781] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 579.808781] env[62569]: DEBUG oslo_concurrency.lockutils [req-ea84cb7f-5d7a-40f7-8cc1-95d94a039b23 req-9f392b4c-44fe-4f24-b6d3-84ae9dbef6f2 service nova] Acquired lock "refresh_cache-ef52ac4a-fa98-4af8-b08a-b34cae526850" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.808781] env[62569]: DEBUG nova.network.neutron [req-ea84cb7f-5d7a-40f7-8cc1-95d94a039b23 req-9f392b4c-44fe-4f24-b6d3-84ae9dbef6f2 service nova] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Refreshing network info cache for port 7f65cf1d-d031-4fe4-bc72-31049bfa7240 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 579.808781] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2b57f27f-8d54-4911-8bda-6d52cebd8dab {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.822125] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89375024-f692-45ef-9f07-5373dfa2a5ce {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.854019] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ef52ac4a-fa98-4af8-b08a-b34cae526850 could not be found. [ 579.854019] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 579.854019] env[62569]: INFO nova.compute.manager [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Took 0.05 seconds to destroy the instance on the hypervisor. [ 579.854019] env[62569]: DEBUG oslo.service.loopingcall [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 579.854019] env[62569]: DEBUG nova.compute.manager [-] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 579.854019] env[62569]: DEBUG nova.network.neutron [-] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 579.885374] env[62569]: DEBUG nova.network.neutron [-] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.027356] env[62569]: DEBUG nova.compute.utils [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 580.031085] env[62569]: DEBUG nova.compute.manager [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 580.031262] env[62569]: DEBUG nova.network.neutron [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 580.135542] env[62569]: DEBUG nova.policy [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e28265bc24244d5fa7e1c9d1f6ae2d26', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '454bd8bed8ed476abe622b2715653f72', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 580.343231] env[62569]: INFO nova.scheduler.client.report [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Deleted allocations for instance 5b1117bd-6210-4df2-b828-f76951752f2f [ 580.354025] env[62569]: DEBUG nova.network.neutron [req-ea84cb7f-5d7a-40f7-8cc1-95d94a039b23 req-9f392b4c-44fe-4f24-b6d3-84ae9dbef6f2 service nova] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.388961] env[62569]: DEBUG nova.network.neutron [-] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.532119] env[62569]: DEBUG nova.compute.manager [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 580.539765] env[62569]: DEBUG nova.network.neutron [req-ea84cb7f-5d7a-40f7-8cc1-95d94a039b23 req-9f392b4c-44fe-4f24-b6d3-84ae9dbef6f2 service nova] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.545019] env[62569]: DEBUG oslo_concurrency.lockutils [None req-03bb0379-2f8b-4bff-a3f2-5046e9c3a5a5 tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.023s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.546388] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.931s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.717304] env[62569]: DEBUG nova.network.neutron [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Successfully created port: c38ec53f-88b5-4901-b088-596a3a85f9c9 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 580.857646] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1a20f2e3-f13a-4fd4-8961-39338eac82dc tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Lock "5b1117bd-6210-4df2-b828-f76951752f2f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.518s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.895146] env[62569]: INFO nova.compute.manager [-] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Took 1.04 seconds to deallocate network for instance. [ 580.897943] env[62569]: DEBUG nova.compute.claims [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 580.898144] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.044562] env[62569]: DEBUG oslo_concurrency.lockutils [req-ea84cb7f-5d7a-40f7-8cc1-95d94a039b23 req-9f392b4c-44fe-4f24-b6d3-84ae9dbef6f2 service nova] Releasing lock "refresh_cache-ef52ac4a-fa98-4af8-b08a-b34cae526850" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.044897] env[62569]: DEBUG nova.compute.manager [req-ea84cb7f-5d7a-40f7-8cc1-95d94a039b23 req-9f392b4c-44fe-4f24-b6d3-84ae9dbef6f2 service nova] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Received event network-vif-deleted-7f65cf1d-d031-4fe4-bc72-31049bfa7240 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 581.148820] env[62569]: DEBUG oslo_concurrency.lockutils [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Acquiring lock "68218523-ae44-4332-8b20-b23559304d60" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.151038] env[62569]: DEBUG oslo_concurrency.lockutils [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Lock "68218523-ae44-4332-8b20-b23559304d60" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.364690] env[62569]: DEBUG nova.compute.manager [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 581.552144] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa8f8a2e-bb8a-4005-b96f-f44093b69c0e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.558103] env[62569]: DEBUG nova.compute.manager [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 581.566883] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6008e035-5f5c-4ac0-8fe5-e031ebfcee16 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.603074] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b0260a-e536-466f-b800-5fcbd54a58a4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.610893] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a51244e-3ce4-4946-83c4-22532e8840af {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.618540] env[62569]: DEBUG nova.virt.hardware [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 581.618822] env[62569]: DEBUG nova.virt.hardware [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 581.619112] env[62569]: DEBUG nova.virt.hardware [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 581.619260] env[62569]: DEBUG nova.virt.hardware [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 581.619480] env[62569]: DEBUG nova.virt.hardware [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 581.619671] env[62569]: DEBUG nova.virt.hardware [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 581.619912] env[62569]: DEBUG nova.virt.hardware [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 581.620128] env[62569]: DEBUG nova.virt.hardware [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 581.620360] env[62569]: DEBUG nova.virt.hardware [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 581.620579] env[62569]: DEBUG nova.virt.hardware [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 581.620802] env[62569]: DEBUG nova.virt.hardware [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 581.621634] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5308ff53-a3b4-4b37-9948-265ff491444c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.635918] env[62569]: DEBUG nova.compute.provider_tree [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.638240] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e2da43-1ff8-456c-ab5e-ff86a57504f8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.903027] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.144174] env[62569]: DEBUG nova.scheduler.client.report [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 582.203022] env[62569]: DEBUG nova.compute.manager [req-d9e0d7b0-92a4-4725-9a2b-1a067aee198d req-c4ce8712-822b-465b-98db-903f38fc47a9 service nova] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Received event network-changed-c38ec53f-88b5-4901-b088-596a3a85f9c9 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 582.203512] env[62569]: DEBUG nova.compute.manager [req-d9e0d7b0-92a4-4725-9a2b-1a067aee198d req-c4ce8712-822b-465b-98db-903f38fc47a9 service nova] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Refreshing instance network info cache due to event network-changed-c38ec53f-88b5-4901-b088-596a3a85f9c9. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 582.203738] env[62569]: DEBUG oslo_concurrency.lockutils [req-d9e0d7b0-92a4-4725-9a2b-1a067aee198d req-c4ce8712-822b-465b-98db-903f38fc47a9 service nova] Acquiring lock "refresh_cache-0dc705ca-6fd1-4844-9c2d-34f2c099ea67" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.203875] env[62569]: DEBUG oslo_concurrency.lockutils [req-d9e0d7b0-92a4-4725-9a2b-1a067aee198d req-c4ce8712-822b-465b-98db-903f38fc47a9 service nova] Acquired lock "refresh_cache-0dc705ca-6fd1-4844-9c2d-34f2c099ea67" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.204209] env[62569]: DEBUG nova.network.neutron [req-d9e0d7b0-92a4-4725-9a2b-1a067aee198d req-c4ce8712-822b-465b-98db-903f38fc47a9 service nova] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Refreshing network info cache for port c38ec53f-88b5-4901-b088-596a3a85f9c9 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 582.307935] env[62569]: ERROR nova.compute.manager [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c38ec53f-88b5-4901-b088-596a3a85f9c9, please check neutron logs for more information. [ 582.307935] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 582.307935] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 582.307935] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 582.307935] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.307935] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 582.307935] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.307935] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 582.307935] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.307935] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 582.307935] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.307935] env[62569]: ERROR nova.compute.manager raise self.value [ 582.307935] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.307935] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 582.307935] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.307935] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 582.308457] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.308457] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 582.308457] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c38ec53f-88b5-4901-b088-596a3a85f9c9, please check neutron logs for more information. [ 582.308457] env[62569]: ERROR nova.compute.manager [ 582.308457] env[62569]: Traceback (most recent call last): [ 582.308457] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 582.308457] env[62569]: listener.cb(fileno) [ 582.308457] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.308457] env[62569]: result = function(*args, **kwargs) [ 582.308457] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.308457] env[62569]: return func(*args, **kwargs) [ 582.308457] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 582.308457] env[62569]: raise e [ 582.308457] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 582.308457] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 582.308457] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.308457] env[62569]: created_port_ids = self._update_ports_for_instance( [ 582.308457] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.308457] env[62569]: with excutils.save_and_reraise_exception(): [ 582.308457] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.308457] env[62569]: self.force_reraise() [ 582.308457] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.308457] env[62569]: raise self.value [ 582.308457] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.308457] env[62569]: updated_port = self._update_port( [ 582.308457] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.308457] env[62569]: _ensure_no_port_binding_failure(port) [ 582.308457] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.308457] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 582.309182] env[62569]: nova.exception.PortBindingFailed: Binding failed for port c38ec53f-88b5-4901-b088-596a3a85f9c9, please check neutron logs for more information. [ 582.309182] env[62569]: Removing descriptor: 17 [ 582.309182] env[62569]: ERROR nova.compute.manager [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c38ec53f-88b5-4901-b088-596a3a85f9c9, please check neutron logs for more information. [ 582.309182] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Traceback (most recent call last): [ 582.309182] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 582.309182] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] yield resources [ 582.309182] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 582.309182] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] self.driver.spawn(context, instance, image_meta, [ 582.309182] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 582.309182] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] self._vmops.spawn(context, instance, image_meta, injected_files, [ 582.309182] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 582.309182] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] vm_ref = self.build_virtual_machine(instance, [ 582.309499] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 582.309499] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] vif_infos = vmwarevif.get_vif_info(self._session, [ 582.309499] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 582.309499] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] for vif in network_info: [ 582.309499] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 582.309499] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] return self._sync_wrapper(fn, *args, **kwargs) [ 582.309499] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 582.309499] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] self.wait() [ 582.309499] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 582.309499] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] self[:] = self._gt.wait() [ 582.309499] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 582.309499] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] return self._exit_event.wait() [ 582.309499] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 582.309812] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] result = hub.switch() [ 582.309812] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 582.309812] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] return self.greenlet.switch() [ 582.309812] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.309812] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] result = function(*args, **kwargs) [ 582.309812] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.309812] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] return func(*args, **kwargs) [ 582.309812] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 582.309812] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] raise e [ 582.309812] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 582.309812] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] nwinfo = self.network_api.allocate_for_instance( [ 582.309812] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.309812] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] created_port_ids = self._update_ports_for_instance( [ 582.310187] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.310187] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] with excutils.save_and_reraise_exception(): [ 582.310187] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.310187] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] self.force_reraise() [ 582.310187] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.310187] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] raise self.value [ 582.310187] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.310187] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] updated_port = self._update_port( [ 582.310187] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.310187] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] _ensure_no_port_binding_failure(port) [ 582.310187] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.310187] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] raise exception.PortBindingFailed(port_id=port['id']) [ 582.310446] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] nova.exception.PortBindingFailed: Binding failed for port c38ec53f-88b5-4901-b088-596a3a85f9c9, please check neutron logs for more information. [ 582.310446] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] [ 582.310446] env[62569]: INFO nova.compute.manager [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Terminating instance [ 582.650072] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.103s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.650667] env[62569]: ERROR nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dcaa61ad-1005-48a6-bbe4-157575908f59, please check neutron logs for more information. [ 582.650667] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Traceback (most recent call last): [ 582.650667] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 582.650667] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] self.driver.spawn(context, instance, image_meta, [ 582.650667] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 582.650667] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 582.650667] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 582.650667] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] vm_ref = self.build_virtual_machine(instance, [ 582.650667] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 582.650667] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] vif_infos = vmwarevif.get_vif_info(self._session, [ 582.650667] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 582.651047] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] for vif in network_info: [ 582.651047] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 582.651047] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] return self._sync_wrapper(fn, *args, **kwargs) [ 582.651047] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 582.651047] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] self.wait() [ 582.651047] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 582.651047] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] self[:] = self._gt.wait() [ 582.651047] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 582.651047] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] return self._exit_event.wait() [ 582.651047] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 582.651047] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] result = hub.switch() [ 582.651047] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 582.651047] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] return self.greenlet.switch() [ 582.651441] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.651441] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] result = function(*args, **kwargs) [ 582.651441] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.651441] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] return func(*args, **kwargs) [ 582.651441] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 582.651441] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] raise e [ 582.651441] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 582.651441] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] nwinfo = self.network_api.allocate_for_instance( [ 582.651441] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.651441] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] created_port_ids = self._update_ports_for_instance( [ 582.651441] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.651441] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] with excutils.save_and_reraise_exception(): [ 582.651441] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.651726] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] self.force_reraise() [ 582.651726] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.651726] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] raise self.value [ 582.651726] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.651726] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] updated_port = self._update_port( [ 582.651726] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.651726] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] _ensure_no_port_binding_failure(port) [ 582.651726] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.651726] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] raise exception.PortBindingFailed(port_id=port['id']) [ 582.651726] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] nova.exception.PortBindingFailed: Binding failed for port dcaa61ad-1005-48a6-bbe4-157575908f59, please check neutron logs for more information. [ 582.651726] env[62569]: ERROR nova.compute.manager [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] [ 582.653221] env[62569]: DEBUG nova.compute.utils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Binding failed for port dcaa61ad-1005-48a6-bbe4-157575908f59, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 582.654478] env[62569]: DEBUG oslo_concurrency.lockutils [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 20.125s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.654656] env[62569]: DEBUG nova.objects.instance [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62569) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 582.657436] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Build of instance 232a126f-37ec-4aea-a602-eb3a13d1d70a was re-scheduled: Binding failed for port dcaa61ad-1005-48a6-bbe4-157575908f59, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 582.657745] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 582.657963] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquiring lock "refresh_cache-232a126f-37ec-4aea-a602-eb3a13d1d70a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.658148] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquired lock "refresh_cache-232a126f-37ec-4aea-a602-eb3a13d1d70a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.658312] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 582.729925] env[62569]: DEBUG nova.network.neutron [req-d9e0d7b0-92a4-4725-9a2b-1a067aee198d req-c4ce8712-822b-465b-98db-903f38fc47a9 service nova] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 582.813856] env[62569]: DEBUG oslo_concurrency.lockutils [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Acquiring lock "refresh_cache-0dc705ca-6fd1-4844-9c2d-34f2c099ea67" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.870527] env[62569]: DEBUG nova.network.neutron [req-d9e0d7b0-92a4-4725-9a2b-1a067aee198d req-c4ce8712-822b-465b-98db-903f38fc47a9 service nova] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.188277] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.300180] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.375988] env[62569]: DEBUG oslo_concurrency.lockutils [req-d9e0d7b0-92a4-4725-9a2b-1a067aee198d req-c4ce8712-822b-465b-98db-903f38fc47a9 service nova] Releasing lock "refresh_cache-0dc705ca-6fd1-4844-9c2d-34f2c099ea67" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.375988] env[62569]: DEBUG oslo_concurrency.lockutils [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Acquired lock "refresh_cache-0dc705ca-6fd1-4844-9c2d-34f2c099ea67" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.376125] env[62569]: DEBUG nova.network.neutron [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 583.480089] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Acquiring lock "311e46eb-85ca-4262-a0a4-21eb073e0dc4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.482100] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Lock "311e46eb-85ca-4262-a0a4-21eb073e0dc4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.670360] env[62569]: DEBUG oslo_concurrency.lockutils [None req-98b5dc8b-42d5-4801-86e8-7b264b108a13 tempest-ServersAdmin275Test-1402741504 tempest-ServersAdmin275Test-1402741504-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.671116] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.381s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.672205] env[62569]: INFO nova.compute.claims [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 583.805336] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Releasing lock "refresh_cache-232a126f-37ec-4aea-a602-eb3a13d1d70a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.805590] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 583.806156] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 583.806156] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 583.829624] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.911594] env[62569]: DEBUG nova.network.neutron [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.002967] env[62569]: DEBUG nova.network.neutron [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.333932] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.351174] env[62569]: DEBUG nova.compute.manager [req-c23ce2b0-d2b4-484a-b7de-208512116814 req-684ce2b3-e22a-4644-9818-8d8ba4904b06 service nova] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Received event network-vif-deleted-c38ec53f-88b5-4901-b088-596a3a85f9c9 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 584.506449] env[62569]: DEBUG oslo_concurrency.lockutils [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Releasing lock "refresh_cache-0dc705ca-6fd1-4844-9c2d-34f2c099ea67" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.506856] env[62569]: DEBUG nova.compute.manager [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 584.507068] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 584.507398] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ca759422-5fdd-423d-aa45-a0ce3a29369f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.517379] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecd34fd2-4206-4b29-a251-9aa944df8426 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.539152] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0dc705ca-6fd1-4844-9c2d-34f2c099ea67 could not be found. [ 584.539782] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 584.539782] env[62569]: INFO nova.compute.manager [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Took 0.03 seconds to destroy the instance on the hypervisor. [ 584.540032] env[62569]: DEBUG oslo.service.loopingcall [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 584.540254] env[62569]: DEBUG nova.compute.manager [-] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 584.540810] env[62569]: DEBUG nova.network.neutron [-] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 584.558414] env[62569]: DEBUG nova.network.neutron [-] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.790310] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Acquiring lock "17f0965b-edf8-4590-9402-2654d5d19f4e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.792390] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Lock "17f0965b-edf8-4590-9402-2654d5d19f4e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.836370] env[62569]: INFO nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 232a126f-37ec-4aea-a602-eb3a13d1d70a] Took 1.03 seconds to deallocate network for instance. [ 585.062194] env[62569]: DEBUG nova.network.neutron [-] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.175391] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d6a81b-47a2-41c2-9116-b0a31d98536d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.184185] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5707df30-2752-4d01-bf1d-b25c85238a91 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.217832] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4be4af9-273b-417f-bb38-0fedeed60117 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.225906] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd4c94a-e00c-44ee-ac44-1bcc376a4dda {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.239681] env[62569]: DEBUG nova.compute.provider_tree [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 585.565514] env[62569]: INFO nova.compute.manager [-] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Took 1.02 seconds to deallocate network for instance. [ 585.567950] env[62569]: DEBUG nova.compute.claims [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 585.568147] env[62569]: DEBUG oslo_concurrency.lockutils [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.743131] env[62569]: DEBUG nova.scheduler.client.report [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 585.905975] env[62569]: INFO nova.scheduler.client.report [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Deleted allocations for instance 232a126f-37ec-4aea-a602-eb3a13d1d70a [ 586.251020] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.578s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.251020] env[62569]: DEBUG nova.compute.manager [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 586.254204] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.361s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.416591] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Lock "232a126f-37ec-4aea-a602-eb3a13d1d70a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.193s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.755373] env[62569]: DEBUG nova.compute.utils [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 586.760625] env[62569]: DEBUG nova.compute.manager [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 586.760831] env[62569]: DEBUG nova.network.neutron [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 586.832553] env[62569]: DEBUG nova.policy [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7f4cb85930884f69aeff1ebbf6abd775', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ddafa4837ce64bf9aec427795e5f48a2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 586.920477] env[62569]: DEBUG nova.compute.manager [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 587.586381] env[62569]: DEBUG nova.compute.manager [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 587.594065] env[62569]: DEBUG nova.network.neutron [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Successfully created port: d04e832e-960a-463f-a2e4-de42cdcbc602 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 587.612095] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.869211] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89696a54-8ae2-43a8-b06a-5e81b4a1e1c8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.874843] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fc10cc9-41e9-4750-8227-17e2cd8aacd1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.904963] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9525ac8e-1173-4b92-ac2c-fa6a0d9bd0b5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.911908] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfdb7f13-ebe4-47b6-8e0c-e254a1fea523 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.924757] env[62569]: DEBUG nova.compute.provider_tree [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 588.427710] env[62569]: DEBUG nova.scheduler.client.report [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 588.485043] env[62569]: DEBUG nova.compute.manager [req-1505f7f1-bdd4-4ead-bab8-41c0763a53ea req-741f0686-2cd7-48e8-98eb-cd5120c11fdf service nova] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Received event network-changed-d04e832e-960a-463f-a2e4-de42cdcbc602 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 588.485596] env[62569]: DEBUG nova.compute.manager [req-1505f7f1-bdd4-4ead-bab8-41c0763a53ea req-741f0686-2cd7-48e8-98eb-cd5120c11fdf service nova] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Refreshing instance network info cache due to event network-changed-d04e832e-960a-463f-a2e4-de42cdcbc602. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 588.485596] env[62569]: DEBUG oslo_concurrency.lockutils [req-1505f7f1-bdd4-4ead-bab8-41c0763a53ea req-741f0686-2cd7-48e8-98eb-cd5120c11fdf service nova] Acquiring lock "refresh_cache-21af21da-87b1-4a84-8544-f1a88494eb08" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.485596] env[62569]: DEBUG oslo_concurrency.lockutils [req-1505f7f1-bdd4-4ead-bab8-41c0763a53ea req-741f0686-2cd7-48e8-98eb-cd5120c11fdf service nova] Acquired lock "refresh_cache-21af21da-87b1-4a84-8544-f1a88494eb08" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.485785] env[62569]: DEBUG nova.network.neutron [req-1505f7f1-bdd4-4ead-bab8-41c0763a53ea req-741f0686-2cd7-48e8-98eb-cd5120c11fdf service nova] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Refreshing network info cache for port d04e832e-960a-463f-a2e4-de42cdcbc602 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 588.601738] env[62569]: DEBUG nova.compute.manager [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 588.633628] env[62569]: DEBUG nova.virt.hardware [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 588.634220] env[62569]: DEBUG nova.virt.hardware [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 588.634560] env[62569]: DEBUG nova.virt.hardware [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 588.634988] env[62569]: DEBUG nova.virt.hardware [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 588.635299] env[62569]: DEBUG nova.virt.hardware [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 588.635747] env[62569]: DEBUG nova.virt.hardware [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 588.638823] env[62569]: DEBUG nova.virt.hardware [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 588.638823] env[62569]: DEBUG nova.virt.hardware [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 588.638823] env[62569]: DEBUG nova.virt.hardware [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 588.638823] env[62569]: DEBUG nova.virt.hardware [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 588.638823] env[62569]: DEBUG nova.virt.hardware [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 588.639148] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63747811-07e6-4638-b4fc-0b920576273f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.648661] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75804cb4-61fa-4e39-8508-a532a3e5e594 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.814702] env[62569]: ERROR nova.compute.manager [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d04e832e-960a-463f-a2e4-de42cdcbc602, please check neutron logs for more information. [ 588.814702] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 588.814702] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 588.814702] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 588.814702] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.814702] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 588.814702] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.814702] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 588.814702] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.814702] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 588.814702] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.814702] env[62569]: ERROR nova.compute.manager raise self.value [ 588.814702] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.814702] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 588.814702] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.814702] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 588.815179] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.815179] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 588.815179] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d04e832e-960a-463f-a2e4-de42cdcbc602, please check neutron logs for more information. [ 588.815179] env[62569]: ERROR nova.compute.manager [ 588.815992] env[62569]: Traceback (most recent call last): [ 588.816113] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 588.816113] env[62569]: listener.cb(fileno) [ 588.816388] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.816388] env[62569]: result = function(*args, **kwargs) [ 588.817134] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 588.817134] env[62569]: return func(*args, **kwargs) [ 588.817134] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 588.817134] env[62569]: raise e [ 588.817134] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 588.817134] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 588.817134] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.817134] env[62569]: created_port_ids = self._update_ports_for_instance( [ 588.817134] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.817134] env[62569]: with excutils.save_and_reraise_exception(): [ 588.817375] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.817375] env[62569]: self.force_reraise() [ 588.817375] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.817375] env[62569]: raise self.value [ 588.817375] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.817375] env[62569]: updated_port = self._update_port( [ 588.817375] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.817375] env[62569]: _ensure_no_port_binding_failure(port) [ 588.817375] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.817375] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 588.817375] env[62569]: nova.exception.PortBindingFailed: Binding failed for port d04e832e-960a-463f-a2e4-de42cdcbc602, please check neutron logs for more information. [ 588.817375] env[62569]: Removing descriptor: 17 [ 588.818086] env[62569]: ERROR nova.compute.manager [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d04e832e-960a-463f-a2e4-de42cdcbc602, please check neutron logs for more information. [ 588.818086] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Traceback (most recent call last): [ 588.818086] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 588.818086] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] yield resources [ 588.818086] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 588.818086] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] self.driver.spawn(context, instance, image_meta, [ 588.818086] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 588.818086] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] self._vmops.spawn(context, instance, image_meta, injected_files, [ 588.818086] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 588.818086] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] vm_ref = self.build_virtual_machine(instance, [ 588.818086] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 588.821756] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] vif_infos = vmwarevif.get_vif_info(self._session, [ 588.821756] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 588.821756] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] for vif in network_info: [ 588.821756] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 588.821756] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] return self._sync_wrapper(fn, *args, **kwargs) [ 588.821756] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 588.821756] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] self.wait() [ 588.821756] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 588.821756] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] self[:] = self._gt.wait() [ 588.821756] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 588.821756] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] return self._exit_event.wait() [ 588.821756] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 588.821756] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] result = hub.switch() [ 588.822046] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 588.822046] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] return self.greenlet.switch() [ 588.822046] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.822046] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] result = function(*args, **kwargs) [ 588.822046] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 588.822046] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] return func(*args, **kwargs) [ 588.822046] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 588.822046] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] raise e [ 588.822046] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 588.822046] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] nwinfo = self.network_api.allocate_for_instance( [ 588.822046] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.822046] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] created_port_ids = self._update_ports_for_instance( [ 588.822046] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.822320] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] with excutils.save_and_reraise_exception(): [ 588.822320] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.822320] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] self.force_reraise() [ 588.822320] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.822320] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] raise self.value [ 588.822320] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.822320] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] updated_port = self._update_port( [ 588.822320] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.822320] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] _ensure_no_port_binding_failure(port) [ 588.822320] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.822320] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] raise exception.PortBindingFailed(port_id=port['id']) [ 588.822320] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] nova.exception.PortBindingFailed: Binding failed for port d04e832e-960a-463f-a2e4-de42cdcbc602, please check neutron logs for more information. [ 588.822320] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] [ 588.822592] env[62569]: INFO nova.compute.manager [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Terminating instance [ 588.935023] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.679s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.935023] env[62569]: ERROR nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8d126f2e-aa3d-4d1a-992e-287dcf02c111, please check neutron logs for more information. [ 588.935023] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Traceback (most recent call last): [ 588.935023] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 588.935023] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] self.driver.spawn(context, instance, image_meta, [ 588.935023] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 588.935023] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 588.935023] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 588.935023] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] vm_ref = self.build_virtual_machine(instance, [ 588.935534] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 588.935534] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] vif_infos = vmwarevif.get_vif_info(self._session, [ 588.935534] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 588.935534] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] for vif in network_info: [ 588.935534] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 588.935534] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] return self._sync_wrapper(fn, *args, **kwargs) [ 588.935534] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 588.935534] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] self.wait() [ 588.935534] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 588.935534] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] self[:] = self._gt.wait() [ 588.935534] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 588.935534] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] return self._exit_event.wait() [ 588.935534] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 588.935855] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] current.throw(*self._exc) [ 588.935855] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.935855] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] result = function(*args, **kwargs) [ 588.935855] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 588.935855] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] return func(*args, **kwargs) [ 588.935855] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 588.935855] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] raise e [ 588.935855] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 588.935855] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] nwinfo = self.network_api.allocate_for_instance( [ 588.935855] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.935855] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] created_port_ids = self._update_ports_for_instance( [ 588.935855] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.935855] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] with excutils.save_and_reraise_exception(): [ 588.936182] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.936182] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] self.force_reraise() [ 588.936182] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.936182] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] raise self.value [ 588.936182] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.936182] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] updated_port = self._update_port( [ 588.936182] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.936182] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] _ensure_no_port_binding_failure(port) [ 588.936182] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.936182] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] raise exception.PortBindingFailed(port_id=port['id']) [ 588.936182] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] nova.exception.PortBindingFailed: Binding failed for port 8d126f2e-aa3d-4d1a-992e-287dcf02c111, please check neutron logs for more information. [ 588.936182] env[62569]: ERROR nova.compute.manager [instance: 11c53593-cf14-4e46-b546-c90a3184028e] [ 588.936566] env[62569]: DEBUG nova.compute.utils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Binding failed for port 8d126f2e-aa3d-4d1a-992e-287dcf02c111, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 588.937775] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.321s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.941074] env[62569]: DEBUG nova.objects.instance [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Lazy-loading 'resources' on Instance uuid 6e35fe1c-05bc-4dcc-b7ec-e9998a647739 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 588.941074] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Build of instance 11c53593-cf14-4e46-b546-c90a3184028e was re-scheduled: Binding failed for port 8d126f2e-aa3d-4d1a-992e-287dcf02c111, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 588.941074] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 588.941074] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquiring lock "refresh_cache-11c53593-cf14-4e46-b546-c90a3184028e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.942184] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquired lock "refresh_cache-11c53593-cf14-4e46-b546-c90a3184028e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.942184] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 589.008196] env[62569]: DEBUG nova.network.neutron [req-1505f7f1-bdd4-4ead-bab8-41c0763a53ea req-741f0686-2cd7-48e8-98eb-cd5120c11fdf service nova] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.091036] env[62569]: DEBUG nova.network.neutron [req-1505f7f1-bdd4-4ead-bab8-41c0763a53ea req-741f0686-2cd7-48e8-98eb-cd5120c11fdf service nova] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.323755] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Acquiring lock "refresh_cache-21af21da-87b1-4a84-8544-f1a88494eb08" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.464408] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.569986] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.592734] env[62569]: DEBUG oslo_concurrency.lockutils [req-1505f7f1-bdd4-4ead-bab8-41c0763a53ea req-741f0686-2cd7-48e8-98eb-cd5120c11fdf service nova] Releasing lock "refresh_cache-21af21da-87b1-4a84-8544-f1a88494eb08" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.593331] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Acquired lock "refresh_cache-21af21da-87b1-4a84-8544-f1a88494eb08" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.593546] env[62569]: DEBUG nova.network.neutron [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 589.869357] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec16b7f7-4132-4cf8-9dc2-c558d3241ac2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.877381] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0290176a-e213-4c7e-96ae-b4eef01882fa {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.909767] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1d9f0a3-b636-4239-a882-70d0e199021c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.917472] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c20dc883-f38f-4c7c-9fdc-3656c4bfc4b0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.931284] env[62569]: DEBUG nova.compute.provider_tree [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 590.022295] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Acquiring lock "3551a2e8-2f7f-475d-9c5c-2c8e88335bc2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.022529] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Lock "3551a2e8-2f7f-475d-9c5c-2c8e88335bc2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.078049] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Releasing lock "refresh_cache-11c53593-cf14-4e46-b546-c90a3184028e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.078049] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 590.078049] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 590.078049] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 590.092722] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.110409] env[62569]: DEBUG nova.network.neutron [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.195797] env[62569]: DEBUG nova.network.neutron [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.434276] env[62569]: DEBUG nova.scheduler.client.report [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 590.508894] env[62569]: DEBUG nova.compute.manager [req-d18a0261-fe79-44f2-8fdb-f738a283fe9d req-a2609ed3-38c2-4a7a-9d07-2af0950e9a4b service nova] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Received event network-vif-deleted-d04e832e-960a-463f-a2e4-de42cdcbc602 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 590.597581] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.698364] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Releasing lock "refresh_cache-21af21da-87b1-4a84-8544-f1a88494eb08" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.698810] env[62569]: DEBUG nova.compute.manager [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 590.699017] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 590.699354] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6a4704ee-9dd1-4706-b535-e388c2e6be5d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.709054] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68c4aa6a-11fb-4177-96df-a8a69c95e50c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.731454] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 21af21da-87b1-4a84-8544-f1a88494eb08 could not be found. [ 590.731677] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 590.731891] env[62569]: INFO nova.compute.manager [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Took 0.03 seconds to destroy the instance on the hypervisor. [ 590.732146] env[62569]: DEBUG oslo.service.loopingcall [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 590.732357] env[62569]: DEBUG nova.compute.manager [-] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 590.732453] env[62569]: DEBUG nova.network.neutron [-] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 590.747410] env[62569]: DEBUG nova.network.neutron [-] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.939261] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.942127] env[62569]: DEBUG oslo_concurrency.lockutils [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.981s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.943168] env[62569]: INFO nova.compute.claims [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 590.964040] env[62569]: INFO nova.scheduler.client.report [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Deleted allocations for instance 6e35fe1c-05bc-4dcc-b7ec-e9998a647739 [ 591.100957] env[62569]: INFO nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: 11c53593-cf14-4e46-b546-c90a3184028e] Took 1.02 seconds to deallocate network for instance. [ 591.249568] env[62569]: DEBUG nova.network.neutron [-] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.471336] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fad8584e-a33f-42dc-9d02-01f439418dfd tempest-ServersAdmin275Test-1645754079 tempest-ServersAdmin275Test-1645754079-project-member] Lock "6e35fe1c-05bc-4dcc-b7ec-e9998a647739" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.065s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.756174] env[62569]: INFO nova.compute.manager [-] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Took 1.02 seconds to deallocate network for instance. [ 591.756174] env[62569]: DEBUG nova.compute.claims [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 591.756472] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.134251] env[62569]: INFO nova.scheduler.client.report [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Deleted allocations for instance 11c53593-cf14-4e46-b546-c90a3184028e [ 592.365819] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29b66fa5-f046-4c2b-9a51-8f2921d2021b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.373943] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfbcf61b-193f-4cd9-a6d7-c9a128d689bc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.406294] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-578b66bc-e0f2-470c-8639-2731caac5be3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.413601] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0198ec08-3c3c-45d3-a1a5-6b0d11d66340 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.426828] env[62569]: DEBUG nova.compute.provider_tree [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.643632] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Lock "11c53593-cf14-4e46-b546-c90a3184028e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.368s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.931487] env[62569]: DEBUG nova.scheduler.client.report [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 593.146818] env[62569]: DEBUG nova.compute.manager [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 593.438351] env[62569]: DEBUG oslo_concurrency.lockutils [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.497s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.438872] env[62569]: DEBUG nova.compute.manager [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 593.444036] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.749s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.454021] env[62569]: DEBUG oslo_concurrency.lockutils [None req-25d13819-d873-47f0-ad48-a6c3caa61677 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquiring lock "a9224414-6d20-4d5f-be5d-2b893117f4b8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.676009] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.964093] env[62569]: DEBUG nova.compute.utils [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 593.968034] env[62569]: DEBUG nova.compute.manager [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 593.968034] env[62569]: DEBUG nova.network.neutron [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 594.024796] env[62569]: DEBUG nova.policy [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '030d3999ecab4c74bbfb30e930af37b5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3999b0896dac4bb6b72c7171612edd2f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 594.365020] env[62569]: DEBUG nova.network.neutron [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Successfully created port: 3d835594-9208-4032-88eb-4bab9d6f3450 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 594.472400] env[62569]: DEBUG nova.compute.manager [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 594.523434] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c0e3550-f36e-4de7-a111-d5d04ed5a097 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.536911] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-862ab096-3856-4033-a929-aa0454f84e3f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.587435] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0435d566-f0be-443e-b91c-7fa880d9a6e8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.599121] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e82b7e-ba90-42d2-a7aa-f9c5de130ceb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.621726] env[62569]: DEBUG nova.compute.provider_tree [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.128019] env[62569]: DEBUG nova.scheduler.client.report [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 595.178100] env[62569]: DEBUG nova.compute.manager [req-59e5ad2b-85e9-4ec9-b65c-82a71cb010af req-da7289a7-de50-4a75-9d9b-232645ed5c12 service nova] [instance: c6292246-847f-4466-909b-6139401943af] Received event network-changed-3d835594-9208-4032-88eb-4bab9d6f3450 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 595.178100] env[62569]: DEBUG nova.compute.manager [req-59e5ad2b-85e9-4ec9-b65c-82a71cb010af req-da7289a7-de50-4a75-9d9b-232645ed5c12 service nova] [instance: c6292246-847f-4466-909b-6139401943af] Refreshing instance network info cache due to event network-changed-3d835594-9208-4032-88eb-4bab9d6f3450. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 595.178100] env[62569]: DEBUG oslo_concurrency.lockutils [req-59e5ad2b-85e9-4ec9-b65c-82a71cb010af req-da7289a7-de50-4a75-9d9b-232645ed5c12 service nova] Acquiring lock "refresh_cache-c6292246-847f-4466-909b-6139401943af" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.178100] env[62569]: DEBUG oslo_concurrency.lockutils [req-59e5ad2b-85e9-4ec9-b65c-82a71cb010af req-da7289a7-de50-4a75-9d9b-232645ed5c12 service nova] Acquired lock "refresh_cache-c6292246-847f-4466-909b-6139401943af" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.178100] env[62569]: DEBUG nova.network.neutron [req-59e5ad2b-85e9-4ec9-b65c-82a71cb010af req-da7289a7-de50-4a75-9d9b-232645ed5c12 service nova] [instance: c6292246-847f-4466-909b-6139401943af] Refreshing network info cache for port 3d835594-9208-4032-88eb-4bab9d6f3450 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 595.361240] env[62569]: ERROR nova.compute.manager [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3d835594-9208-4032-88eb-4bab9d6f3450, please check neutron logs for more information. [ 595.361240] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 595.361240] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 595.361240] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 595.361240] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 595.361240] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 595.361240] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 595.361240] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 595.361240] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.361240] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 595.361240] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.361240] env[62569]: ERROR nova.compute.manager raise self.value [ 595.361240] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 595.361240] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 595.361240] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.361240] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 595.362079] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.362079] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 595.362079] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3d835594-9208-4032-88eb-4bab9d6f3450, please check neutron logs for more information. [ 595.362079] env[62569]: ERROR nova.compute.manager [ 595.362079] env[62569]: Traceback (most recent call last): [ 595.362079] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 595.362079] env[62569]: listener.cb(fileno) [ 595.362079] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.362079] env[62569]: result = function(*args, **kwargs) [ 595.362079] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 595.362079] env[62569]: return func(*args, **kwargs) [ 595.362079] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 595.362079] env[62569]: raise e [ 595.362079] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 595.362079] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 595.362079] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 595.362079] env[62569]: created_port_ids = self._update_ports_for_instance( [ 595.362079] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 595.362079] env[62569]: with excutils.save_and_reraise_exception(): [ 595.362079] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.362079] env[62569]: self.force_reraise() [ 595.362079] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.362079] env[62569]: raise self.value [ 595.362079] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 595.362079] env[62569]: updated_port = self._update_port( [ 595.362079] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.362079] env[62569]: _ensure_no_port_binding_failure(port) [ 595.362079] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.362079] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 595.363967] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 3d835594-9208-4032-88eb-4bab9d6f3450, please check neutron logs for more information. [ 595.363967] env[62569]: Removing descriptor: 17 [ 595.495330] env[62569]: DEBUG nova.compute.manager [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 595.526088] env[62569]: DEBUG nova.virt.hardware [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 595.526358] env[62569]: DEBUG nova.virt.hardware [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 595.526514] env[62569]: DEBUG nova.virt.hardware [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 595.526697] env[62569]: DEBUG nova.virt.hardware [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 595.526841] env[62569]: DEBUG nova.virt.hardware [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 595.526982] env[62569]: DEBUG nova.virt.hardware [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 595.529022] env[62569]: DEBUG nova.virt.hardware [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 595.529223] env[62569]: DEBUG nova.virt.hardware [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 595.529423] env[62569]: DEBUG nova.virt.hardware [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 595.529599] env[62569]: DEBUG nova.virt.hardware [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 595.529783] env[62569]: DEBUG nova.virt.hardware [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 595.530699] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-936242bd-3639-4047-b8b6-acbd7e3f4bdd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.538591] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bacd8c6b-9d88-4d0e-9edc-a4cb6e97929b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.554199] env[62569]: ERROR nova.compute.manager [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3d835594-9208-4032-88eb-4bab9d6f3450, please check neutron logs for more information. [ 595.554199] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] Traceback (most recent call last): [ 595.554199] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 595.554199] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] yield resources [ 595.554199] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 595.554199] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] self.driver.spawn(context, instance, image_meta, [ 595.554199] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 595.554199] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] self._vmops.spawn(context, instance, image_meta, injected_files, [ 595.554199] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 595.554199] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] vm_ref = self.build_virtual_machine(instance, [ 595.554199] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 595.554495] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] vif_infos = vmwarevif.get_vif_info(self._session, [ 595.554495] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 595.554495] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] for vif in network_info: [ 595.554495] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 595.554495] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] return self._sync_wrapper(fn, *args, **kwargs) [ 595.554495] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 595.554495] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] self.wait() [ 595.554495] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 595.554495] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] self[:] = self._gt.wait() [ 595.554495] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 595.554495] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] return self._exit_event.wait() [ 595.554495] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 595.554495] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] current.throw(*self._exc) [ 595.554928] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.554928] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] result = function(*args, **kwargs) [ 595.554928] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 595.554928] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] return func(*args, **kwargs) [ 595.554928] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 595.554928] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] raise e [ 595.554928] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 595.554928] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] nwinfo = self.network_api.allocate_for_instance( [ 595.554928] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 595.554928] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] created_port_ids = self._update_ports_for_instance( [ 595.554928] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 595.554928] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] with excutils.save_and_reraise_exception(): [ 595.554928] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.555218] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] self.force_reraise() [ 595.555218] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.555218] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] raise self.value [ 595.555218] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 595.555218] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] updated_port = self._update_port( [ 595.555218] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.555218] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] _ensure_no_port_binding_failure(port) [ 595.555218] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.555218] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] raise exception.PortBindingFailed(port_id=port['id']) [ 595.555218] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] nova.exception.PortBindingFailed: Binding failed for port 3d835594-9208-4032-88eb-4bab9d6f3450, please check neutron logs for more information. [ 595.555218] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] [ 595.555218] env[62569]: INFO nova.compute.manager [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Terminating instance [ 595.636439] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.192s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.637635] env[62569]: ERROR nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port faec1c73-41a9-4fba-9d98-87f46347132f, please check neutron logs for more information. [ 595.637635] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Traceback (most recent call last): [ 595.637635] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 595.637635] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] self.driver.spawn(context, instance, image_meta, [ 595.637635] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 595.637635] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 595.637635] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 595.637635] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] vm_ref = self.build_virtual_machine(instance, [ 595.637635] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 595.637635] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] vif_infos = vmwarevif.get_vif_info(self._session, [ 595.637635] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 595.637976] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] for vif in network_info: [ 595.637976] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 595.637976] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] return self._sync_wrapper(fn, *args, **kwargs) [ 595.637976] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 595.637976] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] self.wait() [ 595.637976] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 595.637976] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] self[:] = self._gt.wait() [ 595.637976] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 595.637976] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] return self._exit_event.wait() [ 595.637976] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 595.637976] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] current.throw(*self._exc) [ 595.637976] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.637976] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] result = function(*args, **kwargs) [ 595.638319] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 595.638319] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] return func(*args, **kwargs) [ 595.638319] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 595.638319] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] raise e [ 595.638319] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 595.638319] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] nwinfo = self.network_api.allocate_for_instance( [ 595.638319] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 595.638319] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] created_port_ids = self._update_ports_for_instance( [ 595.638319] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 595.638319] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] with excutils.save_and_reraise_exception(): [ 595.638319] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.638319] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] self.force_reraise() [ 595.638319] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.638651] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] raise self.value [ 595.638651] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 595.638651] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] updated_port = self._update_port( [ 595.638651] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.638651] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] _ensure_no_port_binding_failure(port) [ 595.638651] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.638651] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] raise exception.PortBindingFailed(port_id=port['id']) [ 595.638651] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] nova.exception.PortBindingFailed: Binding failed for port faec1c73-41a9-4fba-9d98-87f46347132f, please check neutron logs for more information. [ 595.638651] env[62569]: ERROR nova.compute.manager [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] [ 595.638651] env[62569]: DEBUG nova.compute.utils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Binding failed for port faec1c73-41a9-4fba-9d98-87f46347132f, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 595.638916] env[62569]: DEBUG oslo_concurrency.lockutils [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.607s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.641625] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Build of instance a9224414-6d20-4d5f-be5d-2b893117f4b8 was re-scheduled: Binding failed for port faec1c73-41a9-4fba-9d98-87f46347132f, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 595.642099] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 595.642361] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquiring lock "refresh_cache-a9224414-6d20-4d5f-be5d-2b893117f4b8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.642523] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquired lock "refresh_cache-a9224414-6d20-4d5f-be5d-2b893117f4b8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.642716] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 595.696237] env[62569]: DEBUG nova.network.neutron [req-59e5ad2b-85e9-4ec9-b65c-82a71cb010af req-da7289a7-de50-4a75-9d9b-232645ed5c12 service nova] [instance: c6292246-847f-4466-909b-6139401943af] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.778490] env[62569]: DEBUG nova.network.neutron [req-59e5ad2b-85e9-4ec9-b65c-82a71cb010af req-da7289a7-de50-4a75-9d9b-232645ed5c12 service nova] [instance: c6292246-847f-4466-909b-6139401943af] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.059180] env[62569]: DEBUG oslo_concurrency.lockutils [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Acquiring lock "refresh_cache-c6292246-847f-4466-909b-6139401943af" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.167273] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.243204] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.280528] env[62569]: DEBUG oslo_concurrency.lockutils [req-59e5ad2b-85e9-4ec9-b65c-82a71cb010af req-da7289a7-de50-4a75-9d9b-232645ed5c12 service nova] Releasing lock "refresh_cache-c6292246-847f-4466-909b-6139401943af" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.280914] env[62569]: DEBUG oslo_concurrency.lockutils [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Acquired lock "refresh_cache-c6292246-847f-4466-909b-6139401943af" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.281112] env[62569]: DEBUG nova.network.neutron [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 596.513570] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Acquiring lock "0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.513819] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Lock "0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.604938] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a114d44-495d-4456-bd70-f4c18e66a4ef {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.614415] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11977593-70df-4989-816b-389610b62775 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.646945] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-041d43df-08e4-4198-b46a-5b02b3a415c8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.654430] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ed0e5e2-a131-4dda-92d1-09695f745cce {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.670210] env[62569]: DEBUG nova.compute.provider_tree [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.748168] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Releasing lock "refresh_cache-a9224414-6d20-4d5f-be5d-2b893117f4b8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.748381] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 596.748562] env[62569]: DEBUG nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 596.749177] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 596.767552] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.807036] env[62569]: DEBUG nova.network.neutron [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.901467] env[62569]: DEBUG nova.network.neutron [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.173118] env[62569]: DEBUG nova.scheduler.client.report [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 597.202479] env[62569]: DEBUG nova.compute.manager [req-afc2255b-0501-4b00-8894-010d08b32c3d req-67bbcb34-9e0d-45fa-8817-8a76bffcf7f3 service nova] [instance: c6292246-847f-4466-909b-6139401943af] Received event network-vif-deleted-3d835594-9208-4032-88eb-4bab9d6f3450 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 597.271424] env[62569]: DEBUG nova.network.neutron [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.403880] env[62569]: DEBUG oslo_concurrency.lockutils [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Releasing lock "refresh_cache-c6292246-847f-4466-909b-6139401943af" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.404365] env[62569]: DEBUG nova.compute.manager [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 597.404991] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 597.404991] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f990e9e5-7566-483d-8600-540cfff0ff10 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.413933] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5a4ef9a-1666-4807-8488-cb170221434c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.435211] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c6292246-847f-4466-909b-6139401943af could not be found. [ 597.435441] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 597.435619] env[62569]: INFO nova.compute.manager [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Took 0.03 seconds to destroy the instance on the hypervisor. [ 597.435858] env[62569]: DEBUG oslo.service.loopingcall [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 597.436102] env[62569]: DEBUG nova.compute.manager [-] [instance: c6292246-847f-4466-909b-6139401943af] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 597.436198] env[62569]: DEBUG nova.network.neutron [-] [instance: c6292246-847f-4466-909b-6139401943af] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 597.452012] env[62569]: DEBUG nova.network.neutron [-] [instance: c6292246-847f-4466-909b-6139401943af] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.678078] env[62569]: DEBUG oslo_concurrency.lockutils [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.040s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.678713] env[62569]: ERROR nova.compute.manager [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 04c1526d-781c-4556-ad2a-18d15aa064e0, please check neutron logs for more information. [ 597.678713] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Traceback (most recent call last): [ 597.678713] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 597.678713] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] self.driver.spawn(context, instance, image_meta, [ 597.678713] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 597.678713] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.678713] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.678713] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] vm_ref = self.build_virtual_machine(instance, [ 597.678713] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.678713] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.678713] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.679022] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] for vif in network_info: [ 597.679022] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 597.679022] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] return self._sync_wrapper(fn, *args, **kwargs) [ 597.679022] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 597.679022] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] self.wait() [ 597.679022] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 597.679022] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] self[:] = self._gt.wait() [ 597.679022] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.679022] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] return self._exit_event.wait() [ 597.679022] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 597.679022] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] result = hub.switch() [ 597.679022] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 597.679022] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] return self.greenlet.switch() [ 597.679343] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.679343] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] result = function(*args, **kwargs) [ 597.679343] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.679343] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] return func(*args, **kwargs) [ 597.679343] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 597.679343] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] raise e [ 597.679343] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 597.679343] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] nwinfo = self.network_api.allocate_for_instance( [ 597.679343] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.679343] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] created_port_ids = self._update_ports_for_instance( [ 597.679343] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.679343] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] with excutils.save_and_reraise_exception(): [ 597.679343] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.679841] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] self.force_reraise() [ 597.679841] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.679841] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] raise self.value [ 597.679841] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.679841] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] updated_port = self._update_port( [ 597.679841] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.679841] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] _ensure_no_port_binding_failure(port) [ 597.679841] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.679841] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] raise exception.PortBindingFailed(port_id=port['id']) [ 597.679841] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] nova.exception.PortBindingFailed: Binding failed for port 04c1526d-781c-4556-ad2a-18d15aa064e0, please check neutron logs for more information. [ 597.679841] env[62569]: ERROR nova.compute.manager [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] [ 597.680189] env[62569]: DEBUG nova.compute.utils [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Binding failed for port 04c1526d-781c-4556-ad2a-18d15aa064e0, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 597.681325] env[62569]: DEBUG nova.compute.manager [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Build of instance 49fdf371-5278-442b-b8cc-9a0fc22947b4 was re-scheduled: Binding failed for port 04c1526d-781c-4556-ad2a-18d15aa064e0, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 597.681785] env[62569]: DEBUG nova.compute.manager [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 597.681963] env[62569]: DEBUG oslo_concurrency.lockutils [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Acquiring lock "refresh_cache-49fdf371-5278-442b-b8cc-9a0fc22947b4" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.682122] env[62569]: DEBUG oslo_concurrency.lockutils [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Acquired lock "refresh_cache-49fdf371-5278-442b-b8cc-9a0fc22947b4" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.682279] env[62569]: DEBUG nova.network.neutron [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 597.683701] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.497s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.685264] env[62569]: INFO nova.compute.claims [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 597.773994] env[62569]: INFO nova.compute.manager [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Took 1.03 seconds to deallocate network for instance. [ 597.955052] env[62569]: DEBUG nova.network.neutron [-] [instance: c6292246-847f-4466-909b-6139401943af] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.199848] env[62569]: DEBUG nova.network.neutron [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.276628] env[62569]: DEBUG nova.network.neutron [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.457612] env[62569]: INFO nova.compute.manager [-] [instance: c6292246-847f-4466-909b-6139401943af] Took 1.02 seconds to deallocate network for instance. [ 598.460104] env[62569]: DEBUG nova.compute.claims [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 598.460244] env[62569]: DEBUG oslo_concurrency.lockutils [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.780241] env[62569]: DEBUG oslo_concurrency.lockutils [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Releasing lock "refresh_cache-49fdf371-5278-442b-b8cc-9a0fc22947b4" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.780589] env[62569]: DEBUG nova.compute.manager [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 598.780755] env[62569]: DEBUG nova.compute.manager [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 598.780919] env[62569]: DEBUG nova.network.neutron [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 598.804300] env[62569]: DEBUG nova.network.neutron [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.811979] env[62569]: INFO nova.scheduler.client.report [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Deleted allocations for instance a9224414-6d20-4d5f-be5d-2b893117f4b8 [ 599.070873] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-585a649c-c529-4cb5-918d-4ad3a234290f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.077854] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50ab625f-99f3-48e9-b26d-0da88c546b4d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.107129] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b53d1b6-fff1-4571-8bcd-ab0675362e74 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.114405] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e0b0743-f351-426c-8efa-c848ed231b49 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.127015] env[62569]: DEBUG nova.compute.provider_tree [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.306511] env[62569]: DEBUG nova.network.neutron [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.323268] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2b35e9f8-ecb2-45bd-b815-9003dfa50f35 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Lock "a9224414-6d20-4d5f-be5d-2b893117f4b8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.020s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.323565] env[62569]: DEBUG oslo_concurrency.lockutils [None req-25d13819-d873-47f0-ad48-a6c3caa61677 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Lock "a9224414-6d20-4d5f-be5d-2b893117f4b8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 5.872s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.323778] env[62569]: DEBUG oslo_concurrency.lockutils [None req-25d13819-d873-47f0-ad48-a6c3caa61677 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquiring lock "a9224414-6d20-4d5f-be5d-2b893117f4b8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.323975] env[62569]: DEBUG oslo_concurrency.lockutils [None req-25d13819-d873-47f0-ad48-a6c3caa61677 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Lock "a9224414-6d20-4d5f-be5d-2b893117f4b8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.324151] env[62569]: DEBUG oslo_concurrency.lockutils [None req-25d13819-d873-47f0-ad48-a6c3caa61677 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Lock "a9224414-6d20-4d5f-be5d-2b893117f4b8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.327945] env[62569]: INFO nova.compute.manager [None req-25d13819-d873-47f0-ad48-a6c3caa61677 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Terminating instance [ 599.629605] env[62569]: DEBUG nova.scheduler.client.report [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 599.812693] env[62569]: INFO nova.compute.manager [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] [instance: 49fdf371-5278-442b-b8cc-9a0fc22947b4] Took 1.03 seconds to deallocate network for instance. [ 599.826609] env[62569]: DEBUG nova.compute.manager [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 599.832531] env[62569]: DEBUG oslo_concurrency.lockutils [None req-25d13819-d873-47f0-ad48-a6c3caa61677 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquiring lock "refresh_cache-a9224414-6d20-4d5f-be5d-2b893117f4b8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.832531] env[62569]: DEBUG oslo_concurrency.lockutils [None req-25d13819-d873-47f0-ad48-a6c3caa61677 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Acquired lock "refresh_cache-a9224414-6d20-4d5f-be5d-2b893117f4b8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.832531] env[62569]: DEBUG nova.network.neutron [None req-25d13819-d873-47f0-ad48-a6c3caa61677 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 600.140023] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.455s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.140023] env[62569]: DEBUG nova.compute.manager [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 600.143879] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.244s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.352094] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.354776] env[62569]: DEBUG nova.network.neutron [None req-25d13819-d873-47f0-ad48-a6c3caa61677 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 600.446137] env[62569]: DEBUG nova.network.neutron [None req-25d13819-d873-47f0-ad48-a6c3caa61677 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.647445] env[62569]: DEBUG nova.compute.utils [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 600.652468] env[62569]: DEBUG nova.compute.manager [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 600.652652] env[62569]: DEBUG nova.network.neutron [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 600.734029] env[62569]: DEBUG nova.policy [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b2966ed9bdc74a8c8565aff261ff7a04', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2f040fc2a74549928aa95138d82d1758', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 600.844887] env[62569]: INFO nova.scheduler.client.report [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Deleted allocations for instance 49fdf371-5278-442b-b8cc-9a0fc22947b4 [ 600.950374] env[62569]: DEBUG oslo_concurrency.lockutils [None req-25d13819-d873-47f0-ad48-a6c3caa61677 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Releasing lock "refresh_cache-a9224414-6d20-4d5f-be5d-2b893117f4b8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.950853] env[62569]: DEBUG nova.compute.manager [None req-25d13819-d873-47f0-ad48-a6c3caa61677 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 600.951082] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-25d13819-d873-47f0-ad48-a6c3caa61677 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 600.951585] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9f2a4ebf-7c78-4f1f-a08b-b3ac3147bcfa {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.961590] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b59ffba9-f2cd-4879-931a-244eda8a9bb1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.985877] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-25d13819-d873-47f0-ad48-a6c3caa61677 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a9224414-6d20-4d5f-be5d-2b893117f4b8 could not be found. [ 600.988024] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-25d13819-d873-47f0-ad48-a6c3caa61677 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 600.988024] env[62569]: INFO nova.compute.manager [None req-25d13819-d873-47f0-ad48-a6c3caa61677 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 600.988024] env[62569]: DEBUG oslo.service.loopingcall [None req-25d13819-d873-47f0-ad48-a6c3caa61677 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 600.989051] env[62569]: DEBUG nova.compute.manager [-] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 600.989160] env[62569]: DEBUG nova.network.neutron [-] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 601.011337] env[62569]: DEBUG nova.network.neutron [-] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.106058] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7be2e613-4740-4b6b-9adc-e891b6e93eb7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.115132] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92c3576c-a209-4676-88e6-d24206d57264 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.152609] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df8d9047-7513-47c3-9d68-4c6ec08e641e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.155455] env[62569]: DEBUG nova.compute.manager [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 601.163528] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e916570-6f0c-45cf-8679-f7f818193b06 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.176914] env[62569]: DEBUG nova.compute.provider_tree [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.360771] env[62569]: DEBUG oslo_concurrency.lockutils [None req-768f2687-325c-4f7d-a15f-1e0fafd49769 tempest-ServerPasswordTestJSON-87937161 tempest-ServerPasswordTestJSON-87937161-project-member] Lock "49fdf371-5278-442b-b8cc-9a0fc22947b4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.440s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.423678] env[62569]: DEBUG nova.network.neutron [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Successfully created port: 549efce5-858a-41a7-9e2c-acc7844589e7 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 601.515524] env[62569]: DEBUG nova.network.neutron [-] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.681205] env[62569]: DEBUG nova.scheduler.client.report [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 601.863215] env[62569]: DEBUG nova.compute.manager [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 602.018665] env[62569]: INFO nova.compute.manager [-] [instance: a9224414-6d20-4d5f-be5d-2b893117f4b8] Took 1.03 seconds to deallocate network for instance. [ 602.167388] env[62569]: DEBUG nova.compute.manager [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 602.197026] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.054s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.197026] env[62569]: ERROR nova.compute.manager [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7f65cf1d-d031-4fe4-bc72-31049bfa7240, please check neutron logs for more information. [ 602.197026] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Traceback (most recent call last): [ 602.197026] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 602.197026] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] self.driver.spawn(context, instance, image_meta, [ 602.197026] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 602.197026] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.197026] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.197026] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] vm_ref = self.build_virtual_machine(instance, [ 602.197426] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.197426] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.197426] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.197426] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] for vif in network_info: [ 602.197426] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 602.197426] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] return self._sync_wrapper(fn, *args, **kwargs) [ 602.197426] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 602.197426] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] self.wait() [ 602.197426] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 602.197426] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] self[:] = self._gt.wait() [ 602.197426] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.197426] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] return self._exit_event.wait() [ 602.197426] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 602.198144] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] result = hub.switch() [ 602.198144] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 602.198144] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] return self.greenlet.switch() [ 602.198144] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.198144] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] result = function(*args, **kwargs) [ 602.198144] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.198144] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] return func(*args, **kwargs) [ 602.198144] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 602.198144] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] raise e [ 602.198144] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 602.198144] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] nwinfo = self.network_api.allocate_for_instance( [ 602.198144] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 602.198144] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] created_port_ids = self._update_ports_for_instance( [ 602.199755] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 602.199755] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] with excutils.save_and_reraise_exception(): [ 602.199755] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.199755] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] self.force_reraise() [ 602.199755] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.199755] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] raise self.value [ 602.199755] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 602.199755] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] updated_port = self._update_port( [ 602.199755] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.199755] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] _ensure_no_port_binding_failure(port) [ 602.199755] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.199755] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] raise exception.PortBindingFailed(port_id=port['id']) [ 602.200563] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] nova.exception.PortBindingFailed: Binding failed for port 7f65cf1d-d031-4fe4-bc72-31049bfa7240, please check neutron logs for more information. [ 602.200563] env[62569]: ERROR nova.compute.manager [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] [ 602.200563] env[62569]: DEBUG nova.compute.utils [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Binding failed for port 7f65cf1d-d031-4fe4-bc72-31049bfa7240, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 602.200563] env[62569]: DEBUG nova.virt.hardware [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 602.200766] env[62569]: DEBUG nova.virt.hardware [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 602.201078] env[62569]: DEBUG nova.virt.hardware [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 602.201078] env[62569]: DEBUG nova.virt.hardware [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 602.201155] env[62569]: DEBUG nova.virt.hardware [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 602.201455] env[62569]: DEBUG nova.virt.hardware [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 602.201537] env[62569]: DEBUG nova.virt.hardware [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 602.201686] env[62569]: DEBUG nova.virt.hardware [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 602.201972] env[62569]: DEBUG nova.virt.hardware [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 602.202060] env[62569]: DEBUG nova.virt.hardware [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 602.202215] env[62569]: DEBUG nova.virt.hardware [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 602.202768] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.301s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.203894] env[62569]: INFO nova.compute.claims [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 602.206811] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11cbe1a5-b7f3-4f61-be5a-b12699c8c84a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.213327] env[62569]: DEBUG nova.compute.manager [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Build of instance ef52ac4a-fa98-4af8-b08a-b34cae526850 was re-scheduled: Binding failed for port 7f65cf1d-d031-4fe4-bc72-31049bfa7240, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 602.213783] env[62569]: DEBUG nova.compute.manager [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 602.214016] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Acquiring lock "refresh_cache-ef52ac4a-fa98-4af8-b08a-b34cae526850" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.216253] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Acquired lock "refresh_cache-ef52ac4a-fa98-4af8-b08a-b34cae526850" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.216428] env[62569]: DEBUG nova.network.neutron [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 602.229314] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0e3dc58-ac5b-4526-8c41-c1f827905236 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.251403] env[62569]: DEBUG nova.compute.manager [req-8702cd01-9f53-43ec-b5d4-8bf143419967 req-0d0ce70e-732c-48d1-ac5f-69ec66804f19 service nova] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Received event network-changed-549efce5-858a-41a7-9e2c-acc7844589e7 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 602.251403] env[62569]: DEBUG nova.compute.manager [req-8702cd01-9f53-43ec-b5d4-8bf143419967 req-0d0ce70e-732c-48d1-ac5f-69ec66804f19 service nova] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Refreshing instance network info cache due to event network-changed-549efce5-858a-41a7-9e2c-acc7844589e7. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 602.251657] env[62569]: DEBUG oslo_concurrency.lockutils [req-8702cd01-9f53-43ec-b5d4-8bf143419967 req-0d0ce70e-732c-48d1-ac5f-69ec66804f19 service nova] Acquiring lock "refresh_cache-2a61b4cd-a15c-4064-9351-326dd777603b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.251923] env[62569]: DEBUG oslo_concurrency.lockutils [req-8702cd01-9f53-43ec-b5d4-8bf143419967 req-0d0ce70e-732c-48d1-ac5f-69ec66804f19 service nova] Acquired lock "refresh_cache-2a61b4cd-a15c-4064-9351-326dd777603b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.251923] env[62569]: DEBUG nova.network.neutron [req-8702cd01-9f53-43ec-b5d4-8bf143419967 req-0d0ce70e-732c-48d1-ac5f-69ec66804f19 service nova] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Refreshing network info cache for port 549efce5-858a-41a7-9e2c-acc7844589e7 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 602.386171] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.443855] env[62569]: ERROR nova.compute.manager [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 549efce5-858a-41a7-9e2c-acc7844589e7, please check neutron logs for more information. [ 602.443855] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 602.443855] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 602.443855] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 602.443855] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 602.443855] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 602.443855] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 602.443855] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 602.443855] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.443855] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 602.443855] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.443855] env[62569]: ERROR nova.compute.manager raise self.value [ 602.443855] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 602.443855] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 602.443855] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.443855] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 602.444284] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.444284] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 602.444284] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 549efce5-858a-41a7-9e2c-acc7844589e7, please check neutron logs for more information. [ 602.444284] env[62569]: ERROR nova.compute.manager [ 602.444284] env[62569]: Traceback (most recent call last): [ 602.444284] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 602.444284] env[62569]: listener.cb(fileno) [ 602.444284] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.444284] env[62569]: result = function(*args, **kwargs) [ 602.444284] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.444284] env[62569]: return func(*args, **kwargs) [ 602.444284] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 602.444284] env[62569]: raise e [ 602.444284] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 602.444284] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 602.444284] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 602.444284] env[62569]: created_port_ids = self._update_ports_for_instance( [ 602.444284] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 602.444284] env[62569]: with excutils.save_and_reraise_exception(): [ 602.444284] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.444284] env[62569]: self.force_reraise() [ 602.444284] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.444284] env[62569]: raise self.value [ 602.444284] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 602.444284] env[62569]: updated_port = self._update_port( [ 602.444284] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.444284] env[62569]: _ensure_no_port_binding_failure(port) [ 602.444284] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.444284] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 602.444974] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 549efce5-858a-41a7-9e2c-acc7844589e7, please check neutron logs for more information. [ 602.444974] env[62569]: Removing descriptor: 17 [ 602.444974] env[62569]: ERROR nova.compute.manager [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 549efce5-858a-41a7-9e2c-acc7844589e7, please check neutron logs for more information. [ 602.444974] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Traceback (most recent call last): [ 602.444974] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 602.444974] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] yield resources [ 602.444974] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 602.444974] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] self.driver.spawn(context, instance, image_meta, [ 602.444974] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 602.444974] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.444974] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.444974] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] vm_ref = self.build_virtual_machine(instance, [ 602.445299] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.445299] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.445299] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.445299] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] for vif in network_info: [ 602.445299] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 602.445299] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] return self._sync_wrapper(fn, *args, **kwargs) [ 602.445299] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 602.445299] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] self.wait() [ 602.445299] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 602.445299] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] self[:] = self._gt.wait() [ 602.445299] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.445299] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] return self._exit_event.wait() [ 602.445299] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 602.445683] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] result = hub.switch() [ 602.445683] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 602.445683] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] return self.greenlet.switch() [ 602.445683] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.445683] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] result = function(*args, **kwargs) [ 602.445683] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.445683] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] return func(*args, **kwargs) [ 602.445683] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 602.445683] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] raise e [ 602.445683] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 602.445683] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] nwinfo = self.network_api.allocate_for_instance( [ 602.445683] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 602.445683] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] created_port_ids = self._update_ports_for_instance( [ 602.446031] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 602.446031] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] with excutils.save_and_reraise_exception(): [ 602.446031] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.446031] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] self.force_reraise() [ 602.446031] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.446031] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] raise self.value [ 602.446031] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 602.446031] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] updated_port = self._update_port( [ 602.446031] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.446031] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] _ensure_no_port_binding_failure(port) [ 602.446031] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.446031] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] raise exception.PortBindingFailed(port_id=port['id']) [ 602.446338] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] nova.exception.PortBindingFailed: Binding failed for port 549efce5-858a-41a7-9e2c-acc7844589e7, please check neutron logs for more information. [ 602.446338] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] [ 602.446338] env[62569]: INFO nova.compute.manager [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Terminating instance [ 602.745742] env[62569]: DEBUG nova.network.neutron [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.791051] env[62569]: DEBUG nova.network.neutron [req-8702cd01-9f53-43ec-b5d4-8bf143419967 req-0d0ce70e-732c-48d1-ac5f-69ec66804f19 service nova] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.878403] env[62569]: DEBUG nova.network.neutron [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.894230] env[62569]: DEBUG nova.network.neutron [req-8702cd01-9f53-43ec-b5d4-8bf143419967 req-0d0ce70e-732c-48d1-ac5f-69ec66804f19 service nova] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.949644] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Acquiring lock "refresh_cache-2a61b4cd-a15c-4064-9351-326dd777603b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.051910] env[62569]: DEBUG oslo_concurrency.lockutils [None req-25d13819-d873-47f0-ad48-a6c3caa61677 tempest-ListServersNegativeTestJSON-662261419 tempest-ListServersNegativeTestJSON-662261419-project-member] Lock "a9224414-6d20-4d5f-be5d-2b893117f4b8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.728s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.382545] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Releasing lock "refresh_cache-ef52ac4a-fa98-4af8-b08a-b34cae526850" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.382793] env[62569]: DEBUG nova.compute.manager [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 603.382962] env[62569]: DEBUG nova.compute.manager [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 603.383145] env[62569]: DEBUG nova.network.neutron [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 603.399527] env[62569]: DEBUG oslo_concurrency.lockutils [req-8702cd01-9f53-43ec-b5d4-8bf143419967 req-0d0ce70e-732c-48d1-ac5f-69ec66804f19 service nova] Releasing lock "refresh_cache-2a61b4cd-a15c-4064-9351-326dd777603b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.400420] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Acquired lock "refresh_cache-2a61b4cd-a15c-4064-9351-326dd777603b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.400621] env[62569]: DEBUG nova.network.neutron [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 603.420945] env[62569]: DEBUG nova.network.neutron [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.700270] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54dc91b2-6420-4b6f-8974-472748082770 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.711505] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-814d67f0-e038-47d8-8667-ac652cdee3b6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.744223] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc2a5568-dc28-404b-8df0-6efabe11894c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.751767] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c36a8f69-d474-4032-930e-94d0c5beb746 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.765107] env[62569]: DEBUG nova.compute.provider_tree [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.923705] env[62569]: DEBUG nova.network.neutron [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.929796] env[62569]: DEBUG nova.network.neutron [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.020669] env[62569]: DEBUG nova.network.neutron [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.269948] env[62569]: DEBUG nova.scheduler.client.report [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 604.330334] env[62569]: DEBUG nova.compute.manager [req-bff1135f-15f3-4541-801b-c610401ff632 req-31c7e718-02f6-4ab0-af2b-026e166695f2 service nova] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Received event network-vif-deleted-549efce5-858a-41a7-9e2c-acc7844589e7 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 604.430966] env[62569]: INFO nova.compute.manager [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] [instance: ef52ac4a-fa98-4af8-b08a-b34cae526850] Took 1.05 seconds to deallocate network for instance. [ 604.526089] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Releasing lock "refresh_cache-2a61b4cd-a15c-4064-9351-326dd777603b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.526089] env[62569]: DEBUG nova.compute.manager [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 604.526089] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 604.526089] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-67932ca0-a219-4f34-9642-aa3f8f8cb346 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.533443] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b276f93c-5503-4625-a63b-2dd2381f8d30 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.563129] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2a61b4cd-a15c-4064-9351-326dd777603b could not be found. [ 604.563129] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 604.563129] env[62569]: INFO nova.compute.manager [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 604.563129] env[62569]: DEBUG oslo.service.loopingcall [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 604.563970] env[62569]: DEBUG nova.compute.manager [-] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 604.564201] env[62569]: DEBUG nova.network.neutron [-] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 604.583940] env[62569]: DEBUG nova.network.neutron [-] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.778733] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.576s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.779292] env[62569]: DEBUG nova.compute.manager [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 604.782949] env[62569]: DEBUG oslo_concurrency.lockutils [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.215s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.085480] env[62569]: DEBUG nova.network.neutron [-] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.290049] env[62569]: DEBUG nova.compute.utils [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 605.293955] env[62569]: DEBUG nova.compute.manager [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 605.294399] env[62569]: DEBUG nova.network.neutron [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 605.362690] env[62569]: DEBUG nova.policy [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b2966ed9bdc74a8c8565aff261ff7a04', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2f040fc2a74549928aa95138d82d1758', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 605.477312] env[62569]: INFO nova.scheduler.client.report [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Deleted allocations for instance ef52ac4a-fa98-4af8-b08a-b34cae526850 [ 605.590189] env[62569]: INFO nova.compute.manager [-] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Took 1.03 seconds to deallocate network for instance. [ 605.592799] env[62569]: DEBUG nova.compute.claims [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 605.592980] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.622013] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Acquiring lock "93c8445d-689c-4f77-836f-95eafafbb6a2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.622289] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Lock "93c8445d-689c-4f77-836f-95eafafbb6a2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.686840] env[62569]: DEBUG nova.network.neutron [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Successfully created port: 4d5e63a8-f88c-4b57-93a4-cc764ca6a405 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 605.781340] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5cf92a3-4d66-44e6-b1e0-ef06a6230966 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.793021] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c477451-cc5f-46ce-86b0-646dda675206 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.829264] env[62569]: DEBUG nova.compute.manager [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 605.834102] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8070f598-6505-4b0a-808b-d3e401166d2b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.842956] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-950d289c-2b1d-46d9-bf4b-f9f4427190fe {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.858828] env[62569]: DEBUG nova.compute.provider_tree [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.996018] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fc3a41d9-bba3-4f65-a543-f0eeb5e6097a tempest-DeleteServersAdminTestJSON-1240844293 tempest-DeleteServersAdminTestJSON-1240844293-project-member] Lock "ef52ac4a-fa98-4af8-b08a-b34cae526850" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.926s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.362422] env[62569]: DEBUG nova.scheduler.client.report [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 606.500522] env[62569]: DEBUG nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 606.786883] env[62569]: DEBUG nova.compute.manager [req-dcf38aab-30b2-4233-bfb5-81331e498555 req-9c7747ca-c224-4fbd-a91f-d66fc3b37c9c service nova] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Received event network-changed-4d5e63a8-f88c-4b57-93a4-cc764ca6a405 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 606.786883] env[62569]: DEBUG nova.compute.manager [req-dcf38aab-30b2-4233-bfb5-81331e498555 req-9c7747ca-c224-4fbd-a91f-d66fc3b37c9c service nova] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Refreshing instance network info cache due to event network-changed-4d5e63a8-f88c-4b57-93a4-cc764ca6a405. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 606.787076] env[62569]: DEBUG oslo_concurrency.lockutils [req-dcf38aab-30b2-4233-bfb5-81331e498555 req-9c7747ca-c224-4fbd-a91f-d66fc3b37c9c service nova] Acquiring lock "refresh_cache-76bddf93-ca4f-4987-a446-b369801ebd93" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.787495] env[62569]: DEBUG oslo_concurrency.lockutils [req-dcf38aab-30b2-4233-bfb5-81331e498555 req-9c7747ca-c224-4fbd-a91f-d66fc3b37c9c service nova] Acquired lock "refresh_cache-76bddf93-ca4f-4987-a446-b369801ebd93" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.787692] env[62569]: DEBUG nova.network.neutron [req-dcf38aab-30b2-4233-bfb5-81331e498555 req-9c7747ca-c224-4fbd-a91f-d66fc3b37c9c service nova] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Refreshing network info cache for port 4d5e63a8-f88c-4b57-93a4-cc764ca6a405 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 606.847941] env[62569]: DEBUG nova.compute.manager [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 606.852174] env[62569]: ERROR nova.compute.manager [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4d5e63a8-f88c-4b57-93a4-cc764ca6a405, please check neutron logs for more information. [ 606.852174] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 606.852174] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 606.852174] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 606.852174] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.852174] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 606.852174] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.852174] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 606.852174] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.852174] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 606.852174] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.852174] env[62569]: ERROR nova.compute.manager raise self.value [ 606.852174] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.852174] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 606.852174] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.852174] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 606.852721] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.852721] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 606.852721] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4d5e63a8-f88c-4b57-93a4-cc764ca6a405, please check neutron logs for more information. [ 606.852721] env[62569]: ERROR nova.compute.manager [ 606.852721] env[62569]: Traceback (most recent call last): [ 606.852721] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 606.852721] env[62569]: listener.cb(fileno) [ 606.852721] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.852721] env[62569]: result = function(*args, **kwargs) [ 606.852721] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 606.852721] env[62569]: return func(*args, **kwargs) [ 606.852721] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 606.852721] env[62569]: raise e [ 606.852721] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 606.852721] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 606.852721] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.852721] env[62569]: created_port_ids = self._update_ports_for_instance( [ 606.852721] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.852721] env[62569]: with excutils.save_and_reraise_exception(): [ 606.852721] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.852721] env[62569]: self.force_reraise() [ 606.852721] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.852721] env[62569]: raise self.value [ 606.852721] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.852721] env[62569]: updated_port = self._update_port( [ 606.852721] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.852721] env[62569]: _ensure_no_port_binding_failure(port) [ 606.852721] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.852721] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 606.853533] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 4d5e63a8-f88c-4b57-93a4-cc764ca6a405, please check neutron logs for more information. [ 606.853533] env[62569]: Removing descriptor: 14 [ 606.874096] env[62569]: DEBUG oslo_concurrency.lockutils [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.091s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.874716] env[62569]: ERROR nova.compute.manager [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c38ec53f-88b5-4901-b088-596a3a85f9c9, please check neutron logs for more information. [ 606.874716] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Traceback (most recent call last): [ 606.874716] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 606.874716] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] self.driver.spawn(context, instance, image_meta, [ 606.874716] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 606.874716] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] self._vmops.spawn(context, instance, image_meta, injected_files, [ 606.874716] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 606.874716] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] vm_ref = self.build_virtual_machine(instance, [ 606.874716] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 606.874716] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] vif_infos = vmwarevif.get_vif_info(self._session, [ 606.874716] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 606.875065] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] for vif in network_info: [ 606.875065] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 606.875065] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] return self._sync_wrapper(fn, *args, **kwargs) [ 606.875065] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 606.875065] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] self.wait() [ 606.875065] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 606.875065] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] self[:] = self._gt.wait() [ 606.875065] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 606.875065] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] return self._exit_event.wait() [ 606.875065] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 606.875065] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] result = hub.switch() [ 606.875065] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 606.875065] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] return self.greenlet.switch() [ 606.875351] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.875351] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] result = function(*args, **kwargs) [ 606.875351] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 606.875351] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] return func(*args, **kwargs) [ 606.875351] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 606.875351] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] raise e [ 606.875351] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 606.875351] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] nwinfo = self.network_api.allocate_for_instance( [ 606.875351] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.875351] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] created_port_ids = self._update_ports_for_instance( [ 606.875351] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.875351] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] with excutils.save_and_reraise_exception(): [ 606.875351] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.875691] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] self.force_reraise() [ 606.875691] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.875691] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] raise self.value [ 606.875691] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.875691] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] updated_port = self._update_port( [ 606.875691] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.875691] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] _ensure_no_port_binding_failure(port) [ 606.875691] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.875691] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] raise exception.PortBindingFailed(port_id=port['id']) [ 606.875691] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] nova.exception.PortBindingFailed: Binding failed for port c38ec53f-88b5-4901-b088-596a3a85f9c9, please check neutron logs for more information. [ 606.875691] env[62569]: ERROR nova.compute.manager [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] [ 606.875941] env[62569]: DEBUG nova.compute.utils [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Binding failed for port c38ec53f-88b5-4901-b088-596a3a85f9c9, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 606.878769] env[62569]: DEBUG nova.virt.hardware [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 606.878987] env[62569]: DEBUG nova.virt.hardware [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 606.879157] env[62569]: DEBUG nova.virt.hardware [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 606.879341] env[62569]: DEBUG nova.virt.hardware [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 606.879501] env[62569]: DEBUG nova.virt.hardware [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 606.879781] env[62569]: DEBUG nova.virt.hardware [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 606.879950] env[62569]: DEBUG nova.virt.hardware [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 606.880059] env[62569]: DEBUG nova.virt.hardware [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 606.880225] env[62569]: DEBUG nova.virt.hardware [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 606.880387] env[62569]: DEBUG nova.virt.hardware [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 606.880567] env[62569]: DEBUG nova.virt.hardware [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 606.880863] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.269s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.882295] env[62569]: INFO nova.compute.claims [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 606.885357] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4c38169-4f6c-4d01-8038-e5dc5075dc12 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.888247] env[62569]: DEBUG nova.compute.manager [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Build of instance 0dc705ca-6fd1-4844-9c2d-34f2c099ea67 was re-scheduled: Binding failed for port c38ec53f-88b5-4901-b088-596a3a85f9c9, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 606.888690] env[62569]: DEBUG nova.compute.manager [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 606.888911] env[62569]: DEBUG oslo_concurrency.lockutils [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Acquiring lock "refresh_cache-0dc705ca-6fd1-4844-9c2d-34f2c099ea67" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.889069] env[62569]: DEBUG oslo_concurrency.lockutils [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Acquired lock "refresh_cache-0dc705ca-6fd1-4844-9c2d-34f2c099ea67" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.889229] env[62569]: DEBUG nova.network.neutron [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 606.896228] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d8285fb-8c6a-4b8f-9376-c694109515dd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.914298] env[62569]: ERROR nova.compute.manager [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4d5e63a8-f88c-4b57-93a4-cc764ca6a405, please check neutron logs for more information. [ 606.914298] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Traceback (most recent call last): [ 606.914298] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 606.914298] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] yield resources [ 606.914298] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 606.914298] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] self.driver.spawn(context, instance, image_meta, [ 606.914298] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 606.914298] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] self._vmops.spawn(context, instance, image_meta, injected_files, [ 606.914298] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 606.914298] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] vm_ref = self.build_virtual_machine(instance, [ 606.914298] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 606.914660] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] vif_infos = vmwarevif.get_vif_info(self._session, [ 606.914660] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 606.914660] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] for vif in network_info: [ 606.914660] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 606.914660] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] return self._sync_wrapper(fn, *args, **kwargs) [ 606.914660] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 606.914660] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] self.wait() [ 606.914660] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 606.914660] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] self[:] = self._gt.wait() [ 606.914660] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 606.914660] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] return self._exit_event.wait() [ 606.914660] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 606.914660] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] current.throw(*self._exc) [ 606.914987] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.914987] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] result = function(*args, **kwargs) [ 606.914987] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 606.914987] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] return func(*args, **kwargs) [ 606.914987] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 606.914987] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] raise e [ 606.914987] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 606.914987] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] nwinfo = self.network_api.allocate_for_instance( [ 606.914987] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.914987] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] created_port_ids = self._update_ports_for_instance( [ 606.914987] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.914987] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] with excutils.save_and_reraise_exception(): [ 606.914987] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.915946] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] self.force_reraise() [ 606.915946] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.915946] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] raise self.value [ 606.915946] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.915946] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] updated_port = self._update_port( [ 606.915946] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.915946] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] _ensure_no_port_binding_failure(port) [ 606.915946] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.915946] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] raise exception.PortBindingFailed(port_id=port['id']) [ 606.915946] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] nova.exception.PortBindingFailed: Binding failed for port 4d5e63a8-f88c-4b57-93a4-cc764ca6a405, please check neutron logs for more information. [ 606.915946] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] [ 606.915946] env[62569]: INFO nova.compute.manager [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Terminating instance [ 607.033861] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.307512] env[62569]: DEBUG nova.network.neutron [req-dcf38aab-30b2-4233-bfb5-81331e498555 req-9c7747ca-c224-4fbd-a91f-d66fc3b37c9c service nova] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.355843] env[62569]: DEBUG nova.network.neutron [req-dcf38aab-30b2-4233-bfb5-81331e498555 req-9c7747ca-c224-4fbd-a91f-d66fc3b37c9c service nova] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.410255] env[62569]: DEBUG nova.network.neutron [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.419284] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Acquiring lock "refresh_cache-76bddf93-ca4f-4987-a446-b369801ebd93" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.490219] env[62569]: DEBUG nova.network.neutron [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.859580] env[62569]: DEBUG oslo_concurrency.lockutils [req-dcf38aab-30b2-4233-bfb5-81331e498555 req-9c7747ca-c224-4fbd-a91f-d66fc3b37c9c service nova] Releasing lock "refresh_cache-76bddf93-ca4f-4987-a446-b369801ebd93" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.859580] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Acquired lock "refresh_cache-76bddf93-ca4f-4987-a446-b369801ebd93" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.859857] env[62569]: DEBUG nova.network.neutron [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 607.993999] env[62569]: DEBUG oslo_concurrency.lockutils [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Releasing lock "refresh_cache-0dc705ca-6fd1-4844-9c2d-34f2c099ea67" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.994682] env[62569]: DEBUG nova.compute.manager [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 607.994682] env[62569]: DEBUG nova.compute.manager [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 607.994682] env[62569]: DEBUG nova.network.neutron [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 608.017445] env[62569]: DEBUG nova.network.neutron [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.320390] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54646751-0678-43d1-b878-76052a987568 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.329028] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a782b6-43de-4ea0-90d7-e8a3626c89f9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.365097] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12884e57-bfc8-4a77-b7f1-a2baab37a034 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.374877] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b014b42-7376-4bbf-911d-e37937f31861 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.390225] env[62569]: DEBUG nova.compute.provider_tree [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.400245] env[62569]: DEBUG nova.network.neutron [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.476045] env[62569]: DEBUG nova.network.neutron [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.520257] env[62569]: DEBUG nova.network.neutron [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.895511] env[62569]: DEBUG nova.scheduler.client.report [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 608.924055] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Acquiring lock "3f504d5c-3431-4a80-a7d9-e52af01cf24e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.924292] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Lock "3f504d5c-3431-4a80-a7d9-e52af01cf24e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.979542] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Releasing lock "refresh_cache-76bddf93-ca4f-4987-a446-b369801ebd93" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.980017] env[62569]: DEBUG nova.compute.manager [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 608.980217] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 608.980527] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-64d2348e-d867-4ebb-876c-c1e9f11de180 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.989753] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea90ebf0-befe-40fd-95d0-48272fca51e2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.010854] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 76bddf93-ca4f-4987-a446-b369801ebd93 could not be found. [ 609.011089] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 609.011273] env[62569]: INFO nova.compute.manager [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Took 0.03 seconds to destroy the instance on the hypervisor. [ 609.011508] env[62569]: DEBUG oslo.service.loopingcall [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 609.011711] env[62569]: DEBUG nova.compute.manager [-] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 609.011805] env[62569]: DEBUG nova.network.neutron [-] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 609.026098] env[62569]: INFO nova.compute.manager [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] [instance: 0dc705ca-6fd1-4844-9c2d-34f2c099ea67] Took 1.03 seconds to deallocate network for instance. [ 609.028394] env[62569]: DEBUG nova.compute.manager [req-f8eedc5d-7938-4b7c-8af4-f29e2f91cff6 req-2401c23c-4e07-44d0-a726-6756c6c2faa7 service nova] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Received event network-vif-deleted-4d5e63a8-f88c-4b57-93a4-cc764ca6a405 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 609.029091] env[62569]: DEBUG nova.network.neutron [-] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.402053] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.518s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.402053] env[62569]: DEBUG nova.compute.manager [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 609.402849] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.646s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.530922] env[62569]: DEBUG nova.network.neutron [-] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.694691] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "a0b406c3-9466-41bd-9de1-e675cab2ceef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.694923] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "a0b406c3-9466-41bd-9de1-e675cab2ceef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.904547] env[62569]: DEBUG nova.compute.utils [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 609.905908] env[62569]: DEBUG nova.compute.manager [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Not allocating networking since 'none' was specified. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 610.036146] env[62569]: INFO nova.compute.manager [-] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Took 1.02 seconds to deallocate network for instance. [ 610.038836] env[62569]: DEBUG nova.compute.claims [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 610.038836] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.051281] env[62569]: INFO nova.scheduler.client.report [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Deleted allocations for instance 0dc705ca-6fd1-4844-9c2d-34f2c099ea67 [ 610.252924] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5ed0de2-3867-42cb-9f22-2a0636145a0d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.260389] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-386e8069-71e7-49ba-bc18-891ee05278a1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.288791] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0525fe31-6c95-4bbb-be3b-9fb54768a19b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.295389] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088c94d5-3e43-487b-a7c3-c8605d5cf14a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.308102] env[62569]: DEBUG nova.compute.provider_tree [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.342794] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 610.343031] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 610.410407] env[62569]: DEBUG nova.compute.manager [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 610.561792] env[62569]: DEBUG oslo_concurrency.lockutils [None req-185f8e26-a929-4a95-a65d-103e5a991330 tempest-ServerRescueNegativeTestJSON-818657039 tempest-ServerRescueNegativeTestJSON-818657039-project-member] Lock "0dc705ca-6fd1-4844-9c2d-34f2c099ea67" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.106s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.811693] env[62569]: DEBUG nova.scheduler.client.report [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 610.847799] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 610.848440] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Starting heal instance info cache {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 610.848440] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Rebuilding the list of instances to heal {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10312}} [ 611.065110] env[62569]: DEBUG nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 611.316118] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.913s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.316764] env[62569]: ERROR nova.compute.manager [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d04e832e-960a-463f-a2e4-de42cdcbc602, please check neutron logs for more information. [ 611.316764] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Traceback (most recent call last): [ 611.316764] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 611.316764] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] self.driver.spawn(context, instance, image_meta, [ 611.316764] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 611.316764] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.316764] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.316764] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] vm_ref = self.build_virtual_machine(instance, [ 611.316764] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.316764] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.316764] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.317179] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] for vif in network_info: [ 611.317179] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.317179] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] return self._sync_wrapper(fn, *args, **kwargs) [ 611.317179] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.317179] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] self.wait() [ 611.317179] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.317179] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] self[:] = self._gt.wait() [ 611.317179] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.317179] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] return self._exit_event.wait() [ 611.317179] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.317179] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] result = hub.switch() [ 611.317179] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.317179] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] return self.greenlet.switch() [ 611.317657] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.317657] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] result = function(*args, **kwargs) [ 611.317657] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.317657] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] return func(*args, **kwargs) [ 611.317657] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 611.317657] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] raise e [ 611.317657] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 611.317657] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] nwinfo = self.network_api.allocate_for_instance( [ 611.317657] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.317657] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] created_port_ids = self._update_ports_for_instance( [ 611.317657] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.317657] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] with excutils.save_and_reraise_exception(): [ 611.317657] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.318152] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] self.force_reraise() [ 611.318152] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.318152] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] raise self.value [ 611.318152] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.318152] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] updated_port = self._update_port( [ 611.318152] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.318152] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] _ensure_no_port_binding_failure(port) [ 611.318152] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.318152] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] raise exception.PortBindingFailed(port_id=port['id']) [ 611.318152] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] nova.exception.PortBindingFailed: Binding failed for port d04e832e-960a-463f-a2e4-de42cdcbc602, please check neutron logs for more information. [ 611.318152] env[62569]: ERROR nova.compute.manager [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] [ 611.318569] env[62569]: DEBUG nova.compute.utils [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Binding failed for port d04e832e-960a-463f-a2e4-de42cdcbc602, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 611.318716] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.643s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.320176] env[62569]: INFO nova.compute.claims [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 611.323595] env[62569]: DEBUG nova.compute.manager [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Build of instance 21af21da-87b1-4a84-8544-f1a88494eb08 was re-scheduled: Binding failed for port d04e832e-960a-463f-a2e4-de42cdcbc602, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 611.326017] env[62569]: DEBUG nova.compute.manager [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 611.326017] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Acquiring lock "refresh_cache-21af21da-87b1-4a84-8544-f1a88494eb08" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.326017] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Acquired lock "refresh_cache-21af21da-87b1-4a84-8544-f1a88494eb08" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.326017] env[62569]: DEBUG nova.network.neutron [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 611.351961] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: c6292246-847f-4466-909b-6139401943af] Skipping network cache update for instance because it is Building. {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 611.351961] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Skipping network cache update for instance because it is Building. {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 611.351961] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Skipping network cache update for instance because it is Building. {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 611.351961] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Skipping network cache update for instance because it is Building. {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 611.351961] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Didn't find any instances for network info cache update. {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10394}} [ 611.351961] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 611.352265] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 611.352265] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 611.352347] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 611.352435] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 611.352583] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 611.353309] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62569) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 611.353513] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 611.418959] env[62569]: DEBUG nova.compute.manager [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 611.444463] env[62569]: DEBUG nova.virt.hardware [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 611.444752] env[62569]: DEBUG nova.virt.hardware [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 611.444913] env[62569]: DEBUG nova.virt.hardware [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 611.445109] env[62569]: DEBUG nova.virt.hardware [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 611.445260] env[62569]: DEBUG nova.virt.hardware [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 611.445406] env[62569]: DEBUG nova.virt.hardware [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 611.445615] env[62569]: DEBUG nova.virt.hardware [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 611.445797] env[62569]: DEBUG nova.virt.hardware [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 611.445937] env[62569]: DEBUG nova.virt.hardware [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 611.446113] env[62569]: DEBUG nova.virt.hardware [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 611.446295] env[62569]: DEBUG nova.virt.hardware [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 611.447198] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b7f1a0-cddc-434b-bdc2-4ec2531026d9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.455389] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37b9a806-6870-4ae2-a148-67cc77a8cdfc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.469437] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Instance VIF info [] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 611.475763] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Creating folder: Project (88a6cf1e899640c2b6d619fd54dd76f9). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 611.476080] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-80c32e4e-7d0d-4a4d-b681-9a5bf322bbdf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.485922] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Created folder: Project (88a6cf1e899640c2b6d619fd54dd76f9) in parent group-v269330. [ 611.486136] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Creating folder: Instances. Parent ref: group-v269343. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 611.486365] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7faa37ce-3672-480a-bf5d-290e425b5581 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.493894] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Created folder: Instances in parent group-v269343. [ 611.494219] env[62569]: DEBUG oslo.service.loopingcall [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 611.494336] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 611.494525] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-089ff325-4172-456a-bc75-16496454a397 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.511171] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 611.511171] env[62569]: value = "task-1249844" [ 611.511171] env[62569]: _type = "Task" [ 611.511171] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.520766] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249844, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.596053] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.851110] env[62569]: DEBUG nova.network.neutron [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.859319] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.977932] env[62569]: DEBUG nova.network.neutron [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.022454] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249844, 'name': CreateVM_Task, 'duration_secs': 0.266092} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.022622] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 612.023039] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.023207] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.023511] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 612.023759] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da0208f6-d39b-4061-b706-06775f82c9ec {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.028298] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Waiting for the task: (returnval){ [ 612.028298] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5293b548-da42-16a3-e5cd-f28348d36f41" [ 612.028298] env[62569]: _type = "Task" [ 612.028298] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.037236] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5293b548-da42-16a3-e5cd-f28348d36f41, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.480800] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Releasing lock "refresh_cache-21af21da-87b1-4a84-8544-f1a88494eb08" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.481085] env[62569]: DEBUG nova.compute.manager [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 612.481310] env[62569]: DEBUG nova.compute.manager [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 612.481542] env[62569]: DEBUG nova.network.neutron [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 612.501851] env[62569]: DEBUG nova.network.neutron [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.541838] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5293b548-da42-16a3-e5cd-f28348d36f41, 'name': SearchDatastore_Task, 'duration_secs': 0.009424} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.545543] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.545842] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 612.546096] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.546250] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.546429] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 612.547179] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-94a9eac5-0e48-4058-8e59-ea0fbf8602cd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.555019] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 612.555215] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 612.555992] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-edfe628c-8979-4c33-a0c6-3f8f08e7ce3e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.561285] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Waiting for the task: (returnval){ [ 612.561285] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d6467c-8835-4201-cef2-91eba66b5e52" [ 612.561285] env[62569]: _type = "Task" [ 612.561285] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.571228] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d6467c-8835-4201-cef2-91eba66b5e52, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.740073] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61df8dde-c30f-4f6c-8603-95033b891954 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.748522] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1659f23a-68ce-4709-8f07-4392aa9b3ca6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.779342] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b47d05c-58d4-427b-8588-48f1c62b4896 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.787565] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e7c45d-f203-47dd-8c22-72222b65a936 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.803285] env[62569]: DEBUG nova.compute.provider_tree [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.005102] env[62569]: DEBUG nova.network.neutron [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.072654] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d6467c-8835-4201-cef2-91eba66b5e52, 'name': SearchDatastore_Task, 'duration_secs': 0.009759} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.073628] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d3475f1-72ed-4661-8fcb-9984494c0438 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.078935] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Waiting for the task: (returnval){ [ 613.078935] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d3d099-02a9-95c8-6385-1bbf1768143f" [ 613.078935] env[62569]: _type = "Task" [ 613.078935] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.087022] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d3d099-02a9-95c8-6385-1bbf1768143f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.306769] env[62569]: DEBUG nova.scheduler.client.report [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 613.510789] env[62569]: INFO nova.compute.manager [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: 21af21da-87b1-4a84-8544-f1a88494eb08] Took 1.03 seconds to deallocate network for instance. [ 613.590940] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d3d099-02a9-95c8-6385-1bbf1768143f, 'name': SearchDatastore_Task, 'duration_secs': 0.009725} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.590940] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.590940] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 4b9e4f73-4682-4170-b71e-1d3d1d860577/4b9e4f73-4682-4170-b71e-1d3d1d860577.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 613.590940] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7762bf22-f414-4a33-a3d1-9624e3fea4a8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.598715] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Waiting for the task: (returnval){ [ 613.598715] env[62569]: value = "task-1249845" [ 613.598715] env[62569]: _type = "Task" [ 613.598715] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.606236] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249845, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.812234] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.493s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.812783] env[62569]: DEBUG nova.compute.manager [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 613.815765] env[62569]: DEBUG oslo_concurrency.lockutils [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.355s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.929278] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquiring lock "94392d48-223d-4205-9627-25488a468769" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.929629] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lock "94392d48-223d-4205-9627-25488a468769" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.109565] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249845, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.125883] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquiring lock "015140aa-8336-40e2-856c-723277a48f6e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.125883] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lock "015140aa-8336-40e2-856c-723277a48f6e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.324031] env[62569]: DEBUG nova.compute.utils [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 614.327612] env[62569]: DEBUG nova.compute.manager [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 614.327612] env[62569]: DEBUG nova.network.neutron [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 614.375643] env[62569]: DEBUG nova.policy [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a272992a4894805bfb958680a37da8e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b20340a1ce0447cae1bfd8b7c28928f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 614.545929] env[62569]: INFO nova.scheduler.client.report [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Deleted allocations for instance 21af21da-87b1-4a84-8544-f1a88494eb08 [ 614.610458] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249845, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.650270] env[62569]: DEBUG nova.network.neutron [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Successfully created port: b46ffb38-0127-4d26-812f-fb3830415886 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 614.740996] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83a862ce-1fd2-4bb8-b345-cfe4036bd60d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.748967] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0fab2d5-5d9c-4e39-9ff7-4f653b916766 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.780020] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fc4a62b-9b97-47e0-9b89-236fdfb88e64 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.787935] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d3b123-47ec-450f-a1c3-e50b6e8b6c65 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.803192] env[62569]: DEBUG nova.compute.provider_tree [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 614.827576] env[62569]: DEBUG nova.compute.manager [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 615.056771] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca34314f-5a65-4e83-8242-2845d42bfb7b tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Lock "21af21da-87b1-4a84-8544-f1a88494eb08" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.085s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.111192] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249845, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.488926} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.113263] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 4b9e4f73-4682-4170-b71e-1d3d1d860577/4b9e4f73-4682-4170-b71e-1d3d1d860577.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 615.113263] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 615.113263] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cde57dba-37ec-40b2-ac21-af833ad4f568 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.119352] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Waiting for the task: (returnval){ [ 615.119352] env[62569]: value = "task-1249846" [ 615.119352] env[62569]: _type = "Task" [ 615.119352] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.131174] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249846, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.307594] env[62569]: DEBUG nova.scheduler.client.report [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 615.364389] env[62569]: DEBUG nova.compute.manager [req-b833c6be-7841-4eb8-b2d1-3c9c4839389f req-fa32044e-0db0-4ad5-85d0-513abfd862f5 service nova] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Received event network-changed-b46ffb38-0127-4d26-812f-fb3830415886 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 615.365201] env[62569]: DEBUG nova.compute.manager [req-b833c6be-7841-4eb8-b2d1-3c9c4839389f req-fa32044e-0db0-4ad5-85d0-513abfd862f5 service nova] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Refreshing instance network info cache due to event network-changed-b46ffb38-0127-4d26-812f-fb3830415886. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 615.367109] env[62569]: DEBUG oslo_concurrency.lockutils [req-b833c6be-7841-4eb8-b2d1-3c9c4839389f req-fa32044e-0db0-4ad5-85d0-513abfd862f5 service nova] Acquiring lock "refresh_cache-898f69fe-737a-44cb-9c43-965f1f8685f7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.367109] env[62569]: DEBUG oslo_concurrency.lockutils [req-b833c6be-7841-4eb8-b2d1-3c9c4839389f req-fa32044e-0db0-4ad5-85d0-513abfd862f5 service nova] Acquired lock "refresh_cache-898f69fe-737a-44cb-9c43-965f1f8685f7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.367109] env[62569]: DEBUG nova.network.neutron [req-b833c6be-7841-4eb8-b2d1-3c9c4839389f req-fa32044e-0db0-4ad5-85d0-513abfd862f5 service nova] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Refreshing network info cache for port b46ffb38-0127-4d26-812f-fb3830415886 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 615.477219] env[62569]: ERROR nova.compute.manager [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b46ffb38-0127-4d26-812f-fb3830415886, please check neutron logs for more information. [ 615.477219] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 615.477219] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 615.477219] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 615.477219] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 615.477219] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 615.477219] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 615.477219] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 615.477219] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.477219] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 615.477219] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.477219] env[62569]: ERROR nova.compute.manager raise self.value [ 615.477219] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 615.477219] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 615.477219] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.477219] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 615.477832] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.477832] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 615.477832] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b46ffb38-0127-4d26-812f-fb3830415886, please check neutron logs for more information. [ 615.477832] env[62569]: ERROR nova.compute.manager [ 615.477832] env[62569]: Traceback (most recent call last): [ 615.477832] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 615.477832] env[62569]: listener.cb(fileno) [ 615.477832] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.477832] env[62569]: result = function(*args, **kwargs) [ 615.477832] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 615.477832] env[62569]: return func(*args, **kwargs) [ 615.477832] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 615.477832] env[62569]: raise e [ 615.477832] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 615.477832] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 615.477832] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 615.477832] env[62569]: created_port_ids = self._update_ports_for_instance( [ 615.477832] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 615.477832] env[62569]: with excutils.save_and_reraise_exception(): [ 615.477832] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.477832] env[62569]: self.force_reraise() [ 615.477832] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.477832] env[62569]: raise self.value [ 615.477832] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 615.477832] env[62569]: updated_port = self._update_port( [ 615.477832] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.477832] env[62569]: _ensure_no_port_binding_failure(port) [ 615.477832] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.477832] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 615.478603] env[62569]: nova.exception.PortBindingFailed: Binding failed for port b46ffb38-0127-4d26-812f-fb3830415886, please check neutron logs for more information. [ 615.478603] env[62569]: Removing descriptor: 14 [ 615.558979] env[62569]: DEBUG nova.compute.manager [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 615.630189] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249846, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094134} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.630189] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 615.630751] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be3750e8-42d8-41f5-8c21-e0b1dff047f1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.650193] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Reconfiguring VM instance instance-00000019 to attach disk [datastore1] 4b9e4f73-4682-4170-b71e-1d3d1d860577/4b9e4f73-4682-4170-b71e-1d3d1d860577.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 615.650463] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-566d176e-6879-4fba-be30-f5444c290b20 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.669962] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Waiting for the task: (returnval){ [ 615.669962] env[62569]: value = "task-1249847" [ 615.669962] env[62569]: _type = "Task" [ 615.669962] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.677536] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249847, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.812308] env[62569]: DEBUG oslo_concurrency.lockutils [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.996s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.812949] env[62569]: ERROR nova.compute.manager [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3d835594-9208-4032-88eb-4bab9d6f3450, please check neutron logs for more information. [ 615.812949] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] Traceback (most recent call last): [ 615.812949] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 615.812949] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] self.driver.spawn(context, instance, image_meta, [ 615.812949] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 615.812949] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] self._vmops.spawn(context, instance, image_meta, injected_files, [ 615.812949] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 615.812949] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] vm_ref = self.build_virtual_machine(instance, [ 615.812949] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 615.812949] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] vif_infos = vmwarevif.get_vif_info(self._session, [ 615.812949] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 615.813325] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] for vif in network_info: [ 615.813325] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 615.813325] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] return self._sync_wrapper(fn, *args, **kwargs) [ 615.813325] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 615.813325] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] self.wait() [ 615.813325] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 615.813325] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] self[:] = self._gt.wait() [ 615.813325] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 615.813325] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] return self._exit_event.wait() [ 615.813325] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 615.813325] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] current.throw(*self._exc) [ 615.813325] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.813325] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] result = function(*args, **kwargs) [ 615.813658] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 615.813658] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] return func(*args, **kwargs) [ 615.813658] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 615.813658] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] raise e [ 615.813658] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 615.813658] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] nwinfo = self.network_api.allocate_for_instance( [ 615.813658] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 615.813658] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] created_port_ids = self._update_ports_for_instance( [ 615.813658] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 615.813658] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] with excutils.save_and_reraise_exception(): [ 615.813658] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.813658] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] self.force_reraise() [ 615.813658] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.814112] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] raise self.value [ 615.814112] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 615.814112] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] updated_port = self._update_port( [ 615.814112] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.814112] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] _ensure_no_port_binding_failure(port) [ 615.814112] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.814112] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] raise exception.PortBindingFailed(port_id=port['id']) [ 615.814112] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] nova.exception.PortBindingFailed: Binding failed for port 3d835594-9208-4032-88eb-4bab9d6f3450, please check neutron logs for more information. [ 615.814112] env[62569]: ERROR nova.compute.manager [instance: c6292246-847f-4466-909b-6139401943af] [ 615.814112] env[62569]: DEBUG nova.compute.utils [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Binding failed for port 3d835594-9208-4032-88eb-4bab9d6f3450, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 615.814980] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.463s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.816575] env[62569]: INFO nova.compute.claims [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 615.819509] env[62569]: DEBUG nova.compute.manager [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Build of instance c6292246-847f-4466-909b-6139401943af was re-scheduled: Binding failed for port 3d835594-9208-4032-88eb-4bab9d6f3450, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 615.820140] env[62569]: DEBUG nova.compute.manager [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 615.820369] env[62569]: DEBUG oslo_concurrency.lockutils [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Acquiring lock "refresh_cache-c6292246-847f-4466-909b-6139401943af" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.820539] env[62569]: DEBUG oslo_concurrency.lockutils [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Acquired lock "refresh_cache-c6292246-847f-4466-909b-6139401943af" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.820751] env[62569]: DEBUG nova.network.neutron [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 615.838757] env[62569]: DEBUG nova.compute.manager [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 615.865184] env[62569]: DEBUG nova.virt.hardware [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 615.865438] env[62569]: DEBUG nova.virt.hardware [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 615.865597] env[62569]: DEBUG nova.virt.hardware [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 615.865780] env[62569]: DEBUG nova.virt.hardware [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 615.865926] env[62569]: DEBUG nova.virt.hardware [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 615.866085] env[62569]: DEBUG nova.virt.hardware [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 615.866295] env[62569]: DEBUG nova.virt.hardware [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 615.866452] env[62569]: DEBUG nova.virt.hardware [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 615.866624] env[62569]: DEBUG nova.virt.hardware [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 615.866785] env[62569]: DEBUG nova.virt.hardware [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 615.866954] env[62569]: DEBUG nova.virt.hardware [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 615.868075] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302aec3d-dfff-4d95-afad-2b6d5a40599f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.878203] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbdd37ee-db91-44e3-8f79-f186b70cb65b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.891793] env[62569]: ERROR nova.compute.manager [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b46ffb38-0127-4d26-812f-fb3830415886, please check neutron logs for more information. [ 615.891793] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Traceback (most recent call last): [ 615.891793] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 615.891793] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] yield resources [ 615.891793] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 615.891793] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] self.driver.spawn(context, instance, image_meta, [ 615.891793] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 615.891793] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 615.891793] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 615.891793] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] vm_ref = self.build_virtual_machine(instance, [ 615.891793] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 615.892113] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] vif_infos = vmwarevif.get_vif_info(self._session, [ 615.892113] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 615.892113] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] for vif in network_info: [ 615.892113] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 615.892113] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] return self._sync_wrapper(fn, *args, **kwargs) [ 615.892113] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 615.892113] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] self.wait() [ 615.892113] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 615.892113] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] self[:] = self._gt.wait() [ 615.892113] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 615.892113] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] return self._exit_event.wait() [ 615.892113] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 615.892113] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] current.throw(*self._exc) [ 615.892476] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.892476] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] result = function(*args, **kwargs) [ 615.892476] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 615.892476] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] return func(*args, **kwargs) [ 615.892476] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 615.892476] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] raise e [ 615.892476] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 615.892476] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] nwinfo = self.network_api.allocate_for_instance( [ 615.892476] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 615.892476] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] created_port_ids = self._update_ports_for_instance( [ 615.892476] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 615.892476] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] with excutils.save_and_reraise_exception(): [ 615.892476] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.892814] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] self.force_reraise() [ 615.892814] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.892814] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] raise self.value [ 615.892814] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 615.892814] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] updated_port = self._update_port( [ 615.892814] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.892814] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] _ensure_no_port_binding_failure(port) [ 615.892814] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.892814] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] raise exception.PortBindingFailed(port_id=port['id']) [ 615.892814] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] nova.exception.PortBindingFailed: Binding failed for port b46ffb38-0127-4d26-812f-fb3830415886, please check neutron logs for more information. [ 615.892814] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] [ 615.892814] env[62569]: INFO nova.compute.manager [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Terminating instance [ 615.894088] env[62569]: DEBUG nova.network.neutron [req-b833c6be-7841-4eb8-b2d1-3c9c4839389f req-fa32044e-0db0-4ad5-85d0-513abfd862f5 service nova] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.979253] env[62569]: DEBUG nova.network.neutron [req-b833c6be-7841-4eb8-b2d1-3c9c4839389f req-fa32044e-0db0-4ad5-85d0-513abfd862f5 service nova] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.083968] env[62569]: DEBUG oslo_concurrency.lockutils [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.183746] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249847, 'name': ReconfigVM_Task, 'duration_secs': 0.274493} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.183746] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Reconfigured VM instance instance-00000019 to attach disk [datastore1] 4b9e4f73-4682-4170-b71e-1d3d1d860577/4b9e4f73-4682-4170-b71e-1d3d1d860577.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 616.183746] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c24685ae-eef7-4a9f-9af6-f3131ce117ab {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.190018] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Waiting for the task: (returnval){ [ 616.190018] env[62569]: value = "task-1249848" [ 616.190018] env[62569]: _type = "Task" [ 616.190018] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.195598] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249848, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.344024] env[62569]: DEBUG nova.network.neutron [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.398514] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "refresh_cache-898f69fe-737a-44cb-9c43-965f1f8685f7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.413171] env[62569]: DEBUG nova.network.neutron [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.482349] env[62569]: DEBUG oslo_concurrency.lockutils [req-b833c6be-7841-4eb8-b2d1-3c9c4839389f req-fa32044e-0db0-4ad5-85d0-513abfd862f5 service nova] Releasing lock "refresh_cache-898f69fe-737a-44cb-9c43-965f1f8685f7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.482856] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquired lock "refresh_cache-898f69fe-737a-44cb-9c43-965f1f8685f7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.482945] env[62569]: DEBUG nova.network.neutron [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 616.698717] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249848, 'name': Rename_Task, 'duration_secs': 0.129206} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.699015] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 616.699284] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-709cc705-8db2-4848-b777-a4209c4dbf46 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.705671] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Waiting for the task: (returnval){ [ 616.705671] env[62569]: value = "task-1249849" [ 616.705671] env[62569]: _type = "Task" [ 616.705671] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.713334] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249849, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.920203] env[62569]: DEBUG oslo_concurrency.lockutils [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Releasing lock "refresh_cache-c6292246-847f-4466-909b-6139401943af" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.920490] env[62569]: DEBUG nova.compute.manager [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 616.920692] env[62569]: DEBUG nova.compute.manager [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 616.920865] env[62569]: DEBUG nova.network.neutron [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 616.937940] env[62569]: DEBUG nova.network.neutron [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.002371] env[62569]: DEBUG nova.network.neutron [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.081635] env[62569]: DEBUG nova.network.neutron [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.217720] env[62569]: DEBUG oslo_vmware.api [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249849, 'name': PowerOnVM_Task, 'duration_secs': 0.425422} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.220130] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 617.220403] env[62569]: INFO nova.compute.manager [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Took 5.80 seconds to spawn the instance on the hypervisor. [ 617.220644] env[62569]: DEBUG nova.compute.manager [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 617.221633] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29c2ed39-1ccd-416d-996e-dd2278029e46 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.231918] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed4ca606-56ed-4323-b015-3629492a5124 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.238464] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc766eb8-1bd9-414e-8a2a-76655c211953 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.269333] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47dec29b-49c5-4e50-8c9e-8ba65487157f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.276936] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb1f6c29-40db-4d3d-95e5-3cee3282c3bb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.290339] env[62569]: DEBUG nova.compute.provider_tree [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 617.410068] env[62569]: DEBUG nova.compute.manager [req-345b3e27-5eb2-454d-b10f-202bc7556f01 req-9300a3db-aa1d-43a4-9d01-b2d3a275ce8d service nova] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Received event network-vif-deleted-b46ffb38-0127-4d26-812f-fb3830415886 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 617.440329] env[62569]: DEBUG nova.network.neutron [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.585820] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Releasing lock "refresh_cache-898f69fe-737a-44cb-9c43-965f1f8685f7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.586337] env[62569]: DEBUG nova.compute.manager [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 617.586563] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 617.586907] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-088fe98b-5678-4ddc-af73-cf4be0edb359 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.596096] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-911c8645-112d-4a1a-bf2c-f501a2189889 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.617355] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 898f69fe-737a-44cb-9c43-965f1f8685f7 could not be found. [ 617.617674] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 617.617782] env[62569]: INFO nova.compute.manager [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Took 0.03 seconds to destroy the instance on the hypervisor. [ 617.617967] env[62569]: DEBUG oslo.service.loopingcall [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 617.618203] env[62569]: DEBUG nova.compute.manager [-] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 617.618300] env[62569]: DEBUG nova.network.neutron [-] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 617.634799] env[62569]: DEBUG nova.network.neutron [-] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.739198] env[62569]: INFO nova.compute.manager [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Took 30.15 seconds to build instance. [ 617.792996] env[62569]: DEBUG nova.scheduler.client.report [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 617.944029] env[62569]: INFO nova.compute.manager [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] [instance: c6292246-847f-4466-909b-6139401943af] Took 1.02 seconds to deallocate network for instance. [ 618.138382] env[62569]: DEBUG nova.network.neutron [-] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.242951] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ce458c0d-ea52-410c-84cf-099e9a1d7b97 tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Lock "4b9e4f73-4682-4170-b71e-1d3d1d860577" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.230s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.301135] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.301135] env[62569]: DEBUG nova.compute.manager [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 618.301897] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.916s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.305415] env[62569]: INFO nova.compute.claims [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 618.641811] env[62569]: INFO nova.compute.manager [-] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Took 1.02 seconds to deallocate network for instance. [ 618.643963] env[62569]: DEBUG nova.compute.claims [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 618.644158] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.744696] env[62569]: DEBUG nova.compute.manager [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 618.805552] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Acquiring lock "c515d85e-bcb5-4bac-bacb-1e558f38171f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.805650] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Lock "c515d85e-bcb5-4bac-bacb-1e558f38171f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.810215] env[62569]: DEBUG nova.compute.utils [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 618.815048] env[62569]: DEBUG nova.compute.manager [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 618.815048] env[62569]: DEBUG nova.network.neutron [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 618.865685] env[62569]: DEBUG nova.policy [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '699560fe93eb414ab6dd0ca0e28243b0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '32999948b633402cb38352a6742f1a45', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 618.962725] env[62569]: INFO nova.compute.manager [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Rebuilding instance [ 618.976414] env[62569]: INFO nova.scheduler.client.report [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Deleted allocations for instance c6292246-847f-4466-909b-6139401943af [ 619.012080] env[62569]: DEBUG nova.compute.manager [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 619.013278] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4ec6c47-f78e-4f20-b879-ed73cc5ebb1d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.155495] env[62569]: DEBUG nova.network.neutron [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Successfully created port: a5315c23-9de0-4015-ae25-df72c2a8557e {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 619.269033] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.313101] env[62569]: DEBUG nova.compute.manager [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 619.492419] env[62569]: DEBUG oslo_concurrency.lockutils [None req-18173372-7ae7-429c-8edd-c6a691e507ed tempest-ServerActionsTestOtherA-1308457458 tempest-ServerActionsTestOtherA-1308457458-project-member] Lock "c6292246-847f-4466-909b-6139401943af" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.732688] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc6b171-68bc-44b6-b461-95c7abdb4214 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.740189] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4083ec7b-b222-4061-983c-444fcb798dd2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.773166] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36eb5493-3a01-466c-ac81-de856c46b73d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.780771] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21df12ff-1615-4a4c-8932-f54784a3e700 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.793937] env[62569]: DEBUG nova.compute.provider_tree [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 619.820831] env[62569]: INFO nova.virt.block_device [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Booting with volume a9be157f-b09a-4c72-b5d6-f3dad2ef3936 at /dev/sda [ 619.875421] env[62569]: DEBUG nova.compute.manager [req-9f0e5177-f558-4f66-8914-e3c6575bbcfd req-b42a4f7b-03a5-4da2-9f05-e8e8b38d1522 service nova] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Received event network-changed-a5315c23-9de0-4015-ae25-df72c2a8557e {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 619.875632] env[62569]: DEBUG nova.compute.manager [req-9f0e5177-f558-4f66-8914-e3c6575bbcfd req-b42a4f7b-03a5-4da2-9f05-e8e8b38d1522 service nova] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Refreshing instance network info cache due to event network-changed-a5315c23-9de0-4015-ae25-df72c2a8557e. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 619.876436] env[62569]: DEBUG oslo_concurrency.lockutils [req-9f0e5177-f558-4f66-8914-e3c6575bbcfd req-b42a4f7b-03a5-4da2-9f05-e8e8b38d1522 service nova] Acquiring lock "refresh_cache-94818059-cf10-4dca-a711-fd5809583ce9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.876549] env[62569]: DEBUG oslo_concurrency.lockutils [req-9f0e5177-f558-4f66-8914-e3c6575bbcfd req-b42a4f7b-03a5-4da2-9f05-e8e8b38d1522 service nova] Acquired lock "refresh_cache-94818059-cf10-4dca-a711-fd5809583ce9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.876661] env[62569]: DEBUG nova.network.neutron [req-9f0e5177-f558-4f66-8914-e3c6575bbcfd req-b42a4f7b-03a5-4da2-9f05-e8e8b38d1522 service nova] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Refreshing network info cache for port a5315c23-9de0-4015-ae25-df72c2a8557e {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 619.888698] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-daf2ca65-1005-409b-bdfa-772014c0cb2f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.898699] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-948228bb-2241-4d74-a13c-29a2c84adaec {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.923472] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-624912b6-967f-4aa0-b6a6-fa53e516a372 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.931427] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3ebd4dc-47eb-45dc-ab58-d608b39aaabb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.953617] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af7aad0-ceff-4447-8918-fb960edc8729 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.959984] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-633d315a-0521-43d7-ab96-e16352b9f6be {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.974278] env[62569]: DEBUG nova.virt.block_device [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Updating existing volume attachment record: 9b36e887-afd5-4ad7-9927-1250ca93cbc6 {{(pid=62569) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 619.995741] env[62569]: DEBUG nova.compute.manager [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 620.034352] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 620.034352] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b92d6ad1-54fe-4759-ba15-513d4c18106a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.039678] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Waiting for the task: (returnval){ [ 620.039678] env[62569]: value = "task-1249850" [ 620.039678] env[62569]: _type = "Task" [ 620.039678] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.051398] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249850, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.051590] env[62569]: ERROR nova.compute.manager [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a5315c23-9de0-4015-ae25-df72c2a8557e, please check neutron logs for more information. [ 620.051590] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 620.051590] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 620.051590] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 620.051590] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.051590] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 620.051590] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.051590] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 620.051590] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.051590] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 620.051590] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.051590] env[62569]: ERROR nova.compute.manager raise self.value [ 620.051590] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.051590] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 620.051590] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.051590] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 620.052048] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.052048] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 620.052048] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a5315c23-9de0-4015-ae25-df72c2a8557e, please check neutron logs for more information. [ 620.052048] env[62569]: ERROR nova.compute.manager [ 620.052048] env[62569]: Traceback (most recent call last): [ 620.052048] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 620.052048] env[62569]: listener.cb(fileno) [ 620.052048] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.052048] env[62569]: result = function(*args, **kwargs) [ 620.052048] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 620.052048] env[62569]: return func(*args, **kwargs) [ 620.052048] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 620.052048] env[62569]: raise e [ 620.052048] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 620.052048] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 620.052048] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.052048] env[62569]: created_port_ids = self._update_ports_for_instance( [ 620.052048] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.052048] env[62569]: with excutils.save_and_reraise_exception(): [ 620.052048] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.052048] env[62569]: self.force_reraise() [ 620.052048] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.052048] env[62569]: raise self.value [ 620.052048] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.052048] env[62569]: updated_port = self._update_port( [ 620.052048] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.052048] env[62569]: _ensure_no_port_binding_failure(port) [ 620.052048] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.052048] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 620.052835] env[62569]: nova.exception.PortBindingFailed: Binding failed for port a5315c23-9de0-4015-ae25-df72c2a8557e, please check neutron logs for more information. [ 620.052835] env[62569]: Removing descriptor: 14 [ 620.296901] env[62569]: DEBUG nova.scheduler.client.report [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 620.394747] env[62569]: DEBUG nova.network.neutron [req-9f0e5177-f558-4f66-8914-e3c6575bbcfd req-b42a4f7b-03a5-4da2-9f05-e8e8b38d1522 service nova] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.472383] env[62569]: DEBUG nova.network.neutron [req-9f0e5177-f558-4f66-8914-e3c6575bbcfd req-b42a4f7b-03a5-4da2-9f05-e8e8b38d1522 service nova] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.518189] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.549805] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249850, 'name': PowerOffVM_Task, 'duration_secs': 0.194193} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.550084] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 620.550385] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 620.551089] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dbef7f4-6941-41de-9eea-f4d596991e0a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.557195] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 620.557411] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-562f852f-ca77-4323-8617-16122e4280f6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.584444] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 620.584667] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 620.584846] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Deleting the datastore file [datastore1] 4b9e4f73-4682-4170-b71e-1d3d1d860577 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 620.585124] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b3028da4-5cec-4a5a-a923-90b158f5290a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.591269] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Waiting for the task: (returnval){ [ 620.591269] env[62569]: value = "task-1249852" [ 620.591269] env[62569]: _type = "Task" [ 620.591269] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.599159] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249852, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.802025] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.500s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.802574] env[62569]: DEBUG nova.compute.manager [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 620.805278] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.212s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.974701] env[62569]: DEBUG oslo_concurrency.lockutils [req-9f0e5177-f558-4f66-8914-e3c6575bbcfd req-b42a4f7b-03a5-4da2-9f05-e8e8b38d1522 service nova] Releasing lock "refresh_cache-94818059-cf10-4dca-a711-fd5809583ce9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.100577] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249852, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.095893} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.100867] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 621.101071] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 621.101250] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 621.310880] env[62569]: DEBUG nova.compute.utils [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 621.319590] env[62569]: DEBUG nova.compute.manager [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 621.319829] env[62569]: DEBUG nova.network.neutron [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 621.370112] env[62569]: DEBUG nova.policy [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '978371132fa04e699ad189d588be8240', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1e3fe12587e64fdf808a548d4d703672', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 621.775821] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e3acb4-5110-4875-8244-fbc53611eeab {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.785133] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e798c52d-cf4f-4970-b0c8-a1c05cd4b2d2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.819310] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe5bbef1-35a9-4d91-97ab-059c1cb5c1d3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.822369] env[62569]: DEBUG nova.compute.manager [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 621.835881] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b43c24e4-f3fa-4e70-82ef-44b89eb01b45 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.852123] env[62569]: DEBUG nova.compute.provider_tree [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.923866] env[62569]: DEBUG nova.compute.manager [req-a4419914-733b-4874-8d4a-d373943ecf10 req-38f0bb03-a841-420e-9ad1-4f083aae31c6 service nova] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Received event network-vif-deleted-a5315c23-9de0-4015-ae25-df72c2a8557e {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 621.959062] env[62569]: DEBUG nova.network.neutron [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Successfully created port: 78e8ea53-39d8-4665-a85e-497722b2d3c3 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 622.082017] env[62569]: DEBUG nova.compute.manager [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 622.082596] env[62569]: DEBUG nova.virt.hardware [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 622.082834] env[62569]: DEBUG nova.virt.hardware [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 622.083010] env[62569]: DEBUG nova.virt.hardware [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 622.083198] env[62569]: DEBUG nova.virt.hardware [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 622.083341] env[62569]: DEBUG nova.virt.hardware [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 622.083483] env[62569]: DEBUG nova.virt.hardware [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 622.083732] env[62569]: DEBUG nova.virt.hardware [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 622.083920] env[62569]: DEBUG nova.virt.hardware [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 622.084101] env[62569]: DEBUG nova.virt.hardware [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 622.084266] env[62569]: DEBUG nova.virt.hardware [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 622.084441] env[62569]: DEBUG nova.virt.hardware [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 622.085302] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d68d68f1-5045-4e98-87b4-b89f2d0799c7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.093668] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf3610c4-cf91-440c-b7bc-19a527228b78 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.107585] env[62569]: ERROR nova.compute.manager [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a5315c23-9de0-4015-ae25-df72c2a8557e, please check neutron logs for more information. [ 622.107585] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Traceback (most recent call last): [ 622.107585] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 622.107585] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] yield resources [ 622.107585] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 622.107585] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] self.driver.spawn(context, instance, image_meta, [ 622.107585] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 622.107585] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 622.107585] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 622.107585] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] vm_ref = self.build_virtual_machine(instance, [ 622.107585] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 622.108085] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] vif_infos = vmwarevif.get_vif_info(self._session, [ 622.108085] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 622.108085] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] for vif in network_info: [ 622.108085] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 622.108085] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] return self._sync_wrapper(fn, *args, **kwargs) [ 622.108085] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 622.108085] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] self.wait() [ 622.108085] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 622.108085] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] self[:] = self._gt.wait() [ 622.108085] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 622.108085] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] return self._exit_event.wait() [ 622.108085] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 622.108085] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] current.throw(*self._exc) [ 622.108498] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.108498] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] result = function(*args, **kwargs) [ 622.108498] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.108498] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] return func(*args, **kwargs) [ 622.108498] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 622.108498] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] raise e [ 622.108498] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 622.108498] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] nwinfo = self.network_api.allocate_for_instance( [ 622.108498] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 622.108498] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] created_port_ids = self._update_ports_for_instance( [ 622.108498] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 622.108498] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] with excutils.save_and_reraise_exception(): [ 622.108498] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.108856] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] self.force_reraise() [ 622.108856] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.108856] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] raise self.value [ 622.108856] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 622.108856] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] updated_port = self._update_port( [ 622.108856] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.108856] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] _ensure_no_port_binding_failure(port) [ 622.108856] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.108856] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] raise exception.PortBindingFailed(port_id=port['id']) [ 622.108856] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] nova.exception.PortBindingFailed: Binding failed for port a5315c23-9de0-4015-ae25-df72c2a8557e, please check neutron logs for more information. [ 622.108856] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] [ 622.108856] env[62569]: INFO nova.compute.manager [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Terminating instance [ 622.135504] env[62569]: DEBUG nova.virt.hardware [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 622.135860] env[62569]: DEBUG nova.virt.hardware [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 622.135953] env[62569]: DEBUG nova.virt.hardware [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 622.136075] env[62569]: DEBUG nova.virt.hardware [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 622.136228] env[62569]: DEBUG nova.virt.hardware [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 622.136375] env[62569]: DEBUG nova.virt.hardware [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 622.136604] env[62569]: DEBUG nova.virt.hardware [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 622.136735] env[62569]: DEBUG nova.virt.hardware [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 622.136899] env[62569]: DEBUG nova.virt.hardware [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 622.137068] env[62569]: DEBUG nova.virt.hardware [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 622.137237] env[62569]: DEBUG nova.virt.hardware [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 622.138284] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aee52df-debc-4656-80c2-f8d12b830429 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.145971] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-759a1ca7-513f-437d-9657-e23a01f5b5a6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.159075] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Instance VIF info [] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 622.164625] env[62569]: DEBUG oslo.service.loopingcall [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 622.164876] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 622.165089] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-023f3fcd-0b7b-4d30-a380-5f27ffacb8dd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.183164] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 622.183164] env[62569]: value = "task-1249853" [ 622.183164] env[62569]: _type = "Task" [ 622.183164] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.191743] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249853, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.355592] env[62569]: DEBUG nova.scheduler.client.report [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 622.611957] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Acquiring lock "refresh_cache-94818059-cf10-4dca-a711-fd5809583ce9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.612187] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Acquired lock "refresh_cache-94818059-cf10-4dca-a711-fd5809583ce9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.612370] env[62569]: DEBUG nova.network.neutron [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 622.696416] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249853, 'name': CreateVM_Task, 'duration_secs': 0.25769} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.697022] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 622.697105] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.697299] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.697744] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 622.698019] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a67d8e41-99fb-4754-9067-794893baf1c7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.704452] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Waiting for the task: (returnval){ [ 622.704452] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52437806-f739-4d86-98e9-af52ec78e2cb" [ 622.704452] env[62569]: _type = "Task" [ 622.704452] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.712457] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52437806-f739-4d86-98e9-af52ec78e2cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.841702] env[62569]: DEBUG nova.compute.manager [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 622.863310] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.056s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.863310] env[62569]: ERROR nova.compute.manager [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 549efce5-858a-41a7-9e2c-acc7844589e7, please check neutron logs for more information. [ 622.863310] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Traceback (most recent call last): [ 622.863310] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 622.863310] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] self.driver.spawn(context, instance, image_meta, [ 622.863310] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 622.863310] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 622.863310] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 622.863310] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] vm_ref = self.build_virtual_machine(instance, [ 622.863670] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 622.863670] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] vif_infos = vmwarevif.get_vif_info(self._session, [ 622.863670] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 622.863670] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] for vif in network_info: [ 622.863670] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 622.863670] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] return self._sync_wrapper(fn, *args, **kwargs) [ 622.863670] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 622.863670] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] self.wait() [ 622.863670] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 622.863670] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] self[:] = self._gt.wait() [ 622.863670] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 622.863670] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] return self._exit_event.wait() [ 622.863670] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 622.865229] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] result = hub.switch() [ 622.865229] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 622.865229] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] return self.greenlet.switch() [ 622.865229] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.865229] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] result = function(*args, **kwargs) [ 622.865229] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.865229] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] return func(*args, **kwargs) [ 622.865229] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 622.865229] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] raise e [ 622.865229] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 622.865229] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] nwinfo = self.network_api.allocate_for_instance( [ 622.865229] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 622.865229] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] created_port_ids = self._update_ports_for_instance( [ 622.865597] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 622.865597] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] with excutils.save_and_reraise_exception(): [ 622.865597] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.865597] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] self.force_reraise() [ 622.865597] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.865597] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] raise self.value [ 622.865597] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 622.865597] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] updated_port = self._update_port( [ 622.865597] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.865597] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] _ensure_no_port_binding_failure(port) [ 622.865597] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.865597] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] raise exception.PortBindingFailed(port_id=port['id']) [ 622.865922] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] nova.exception.PortBindingFailed: Binding failed for port 549efce5-858a-41a7-9e2c-acc7844589e7, please check neutron logs for more information. [ 622.865922] env[62569]: ERROR nova.compute.manager [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] [ 622.865922] env[62569]: DEBUG nova.compute.utils [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Binding failed for port 549efce5-858a-41a7-9e2c-acc7844589e7, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 622.867678] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.834s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.869945] env[62569]: INFO nova.compute.claims [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 622.873359] env[62569]: DEBUG nova.compute.manager [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Build of instance 2a61b4cd-a15c-4064-9351-326dd777603b was re-scheduled: Binding failed for port 549efce5-858a-41a7-9e2c-acc7844589e7, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 622.873824] env[62569]: DEBUG nova.compute.manager [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 622.874070] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Acquiring lock "refresh_cache-2a61b4cd-a15c-4064-9351-326dd777603b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.874211] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Acquired lock "refresh_cache-2a61b4cd-a15c-4064-9351-326dd777603b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.874380] env[62569]: DEBUG nova.network.neutron [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 622.886781] env[62569]: DEBUG nova.virt.hardware [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 622.887126] env[62569]: DEBUG nova.virt.hardware [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 622.887318] env[62569]: DEBUG nova.virt.hardware [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 622.887510] env[62569]: DEBUG nova.virt.hardware [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 622.887768] env[62569]: DEBUG nova.virt.hardware [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 622.887940] env[62569]: DEBUG nova.virt.hardware [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 622.888167] env[62569]: DEBUG nova.virt.hardware [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 622.889064] env[62569]: DEBUG nova.virt.hardware [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 622.889064] env[62569]: DEBUG nova.virt.hardware [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 622.889064] env[62569]: DEBUG nova.virt.hardware [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 622.889513] env[62569]: DEBUG nova.virt.hardware [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 622.891032] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ca2abee-ebe8-41e3-806b-2aa6173bbdbb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.905554] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5966e44f-e845-4d96-8876-d0be5b828b45 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.133496] env[62569]: DEBUG nova.network.neutron [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.144637] env[62569]: ERROR nova.compute.manager [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 78e8ea53-39d8-4665-a85e-497722b2d3c3, please check neutron logs for more information. [ 623.144637] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 623.144637] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 623.144637] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 623.144637] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 623.144637] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 623.144637] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 623.144637] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 623.144637] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.144637] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 623.144637] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.144637] env[62569]: ERROR nova.compute.manager raise self.value [ 623.144637] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 623.144637] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 623.144637] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.144637] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 623.145148] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.145148] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 623.145148] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 78e8ea53-39d8-4665-a85e-497722b2d3c3, please check neutron logs for more information. [ 623.145148] env[62569]: ERROR nova.compute.manager [ 623.145455] env[62569]: Traceback (most recent call last): [ 623.145549] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 623.145549] env[62569]: listener.cb(fileno) [ 623.145623] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.145623] env[62569]: result = function(*args, **kwargs) [ 623.145698] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 623.145698] env[62569]: return func(*args, **kwargs) [ 623.145767] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 623.145767] env[62569]: raise e [ 623.145834] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 623.145834] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 623.145901] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 623.145901] env[62569]: created_port_ids = self._update_ports_for_instance( [ 623.145969] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 623.145969] env[62569]: with excutils.save_and_reraise_exception(): [ 623.146065] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.146065] env[62569]: self.force_reraise() [ 623.146133] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.146133] env[62569]: raise self.value [ 623.146203] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 623.146203] env[62569]: updated_port = self._update_port( [ 623.146798] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.146798] env[62569]: _ensure_no_port_binding_failure(port) [ 623.146798] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.146798] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 623.146798] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 78e8ea53-39d8-4665-a85e-497722b2d3c3, please check neutron logs for more information. [ 623.146798] env[62569]: Removing descriptor: 14 [ 623.147052] env[62569]: ERROR nova.compute.manager [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 78e8ea53-39d8-4665-a85e-497722b2d3c3, please check neutron logs for more information. [ 623.147052] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Traceback (most recent call last): [ 623.147052] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 623.147052] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] yield resources [ 623.147052] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 623.147052] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] self.driver.spawn(context, instance, image_meta, [ 623.147052] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 623.147052] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] self._vmops.spawn(context, instance, image_meta, injected_files, [ 623.147052] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 623.147052] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] vm_ref = self.build_virtual_machine(instance, [ 623.147052] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 623.147369] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] vif_infos = vmwarevif.get_vif_info(self._session, [ 623.147369] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 623.147369] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] for vif in network_info: [ 623.147369] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 623.147369] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] return self._sync_wrapper(fn, *args, **kwargs) [ 623.147369] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 623.147369] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] self.wait() [ 623.147369] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 623.147369] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] self[:] = self._gt.wait() [ 623.147369] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 623.147369] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] return self._exit_event.wait() [ 623.147369] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 623.147369] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] result = hub.switch() [ 623.147732] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 623.147732] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] return self.greenlet.switch() [ 623.147732] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.147732] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] result = function(*args, **kwargs) [ 623.147732] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 623.147732] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] return func(*args, **kwargs) [ 623.147732] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 623.147732] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] raise e [ 623.147732] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 623.147732] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] nwinfo = self.network_api.allocate_for_instance( [ 623.147732] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 623.147732] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] created_port_ids = self._update_ports_for_instance( [ 623.147732] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 623.148102] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] with excutils.save_and_reraise_exception(): [ 623.148102] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.148102] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] self.force_reraise() [ 623.148102] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.148102] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] raise self.value [ 623.148102] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 623.148102] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] updated_port = self._update_port( [ 623.148102] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.148102] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] _ensure_no_port_binding_failure(port) [ 623.148102] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.148102] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] raise exception.PortBindingFailed(port_id=port['id']) [ 623.148102] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] nova.exception.PortBindingFailed: Binding failed for port 78e8ea53-39d8-4665-a85e-497722b2d3c3, please check neutron logs for more information. [ 623.148102] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] [ 623.148459] env[62569]: INFO nova.compute.manager [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Terminating instance [ 623.209500] env[62569]: DEBUG nova.network.neutron [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.216500] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52437806-f739-4d86-98e9-af52ec78e2cb, 'name': SearchDatastore_Task, 'duration_secs': 0.008484} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.216787] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.217024] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 623.217255] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.217400] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.217584] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 623.217837] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1c20036a-2d71-43be-ab5d-a2b6543fd138 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.226308] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 623.226484] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 623.227171] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e3a2e53-5bc9-40cf-9889-1a3f69a1c20e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.232367] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Waiting for the task: (returnval){ [ 623.232367] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52bb7914-ddbd-55fe-3951-b05f66aeebf5" [ 623.232367] env[62569]: _type = "Task" [ 623.232367] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.240028] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52bb7914-ddbd-55fe-3951-b05f66aeebf5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.393968] env[62569]: DEBUG nova.network.neutron [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.440476] env[62569]: DEBUG nova.network.neutron [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.656700] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Acquiring lock "refresh_cache-c29ad19e-91a5-4c3d-912b-bd7be4c42016" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.656915] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Acquired lock "refresh_cache-c29ad19e-91a5-4c3d-912b-bd7be4c42016" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.657110] env[62569]: DEBUG nova.network.neutron [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 623.712288] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Releasing lock "refresh_cache-94818059-cf10-4dca-a711-fd5809583ce9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.712859] env[62569]: DEBUG nova.compute.manager [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 623.713196] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-750759d2-62a2-47ef-bbe0-e71fa17e5f8e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.722088] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b82f5c8-b6a8-46a7-a159-4ef067f1f304 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.745933] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52bb7914-ddbd-55fe-3951-b05f66aeebf5, 'name': SearchDatastore_Task, 'duration_secs': 0.007908} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.751216] env[62569]: WARNING nova.virt.vmwareapi.driver [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 94818059-cf10-4dca-a711-fd5809583ce9 could not be found. [ 623.751422] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 623.751686] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8af3717d-4792-4639-ba8c-b52aa9873453 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.753794] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fbbb4994-4342-460e-8996-a60c0301d179 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.758085] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Waiting for the task: (returnval){ [ 623.758085] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]523dd38c-0e29-cc5a-8aa8-7c535450b9e8" [ 623.758085] env[62569]: _type = "Task" [ 623.758085] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.765519] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ef2983f-4bcc-4402-9d87-21644cfbdb50 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.780891] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]523dd38c-0e29-cc5a-8aa8-7c535450b9e8, 'name': SearchDatastore_Task, 'duration_secs': 0.008319} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.781067] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.781320] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 4b9e4f73-4682-4170-b71e-1d3d1d860577/4b9e4f73-4682-4170-b71e-1d3d1d860577.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 623.781551] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1e38bb01-0f72-4b6a-8f79-0d14d9b67507 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.787404] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 94818059-cf10-4dca-a711-fd5809583ce9 could not be found. [ 623.787640] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 623.787853] env[62569]: INFO nova.compute.manager [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Took 0.07 seconds to destroy the instance on the hypervisor. [ 623.788249] env[62569]: DEBUG oslo.service.loopingcall [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 623.788566] env[62569]: DEBUG nova.compute.manager [-] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 623.788662] env[62569]: DEBUG nova.network.neutron [-] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 623.792900] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Waiting for the task: (returnval){ [ 623.792900] env[62569]: value = "task-1249854" [ 623.792900] env[62569]: _type = "Task" [ 623.792900] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.800406] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249854, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.807207] env[62569]: DEBUG nova.network.neutron [-] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.945802] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Releasing lock "refresh_cache-2a61b4cd-a15c-4064-9351-326dd777603b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.946078] env[62569]: DEBUG nova.compute.manager [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 623.946245] env[62569]: DEBUG nova.compute.manager [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 623.946419] env[62569]: DEBUG nova.network.neutron [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 623.964472] env[62569]: DEBUG nova.network.neutron [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.179629] env[62569]: DEBUG nova.network.neutron [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.196898] env[62569]: DEBUG nova.compute.manager [req-1defd460-b47e-4224-af8c-042b4bd61d86 req-0767f3aa-5f3c-44ed-9859-566ac5992f0b service nova] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Received event network-changed-78e8ea53-39d8-4665-a85e-497722b2d3c3 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 624.197113] env[62569]: DEBUG nova.compute.manager [req-1defd460-b47e-4224-af8c-042b4bd61d86 req-0767f3aa-5f3c-44ed-9859-566ac5992f0b service nova] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Refreshing instance network info cache due to event network-changed-78e8ea53-39d8-4665-a85e-497722b2d3c3. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 624.197313] env[62569]: DEBUG oslo_concurrency.lockutils [req-1defd460-b47e-4224-af8c-042b4bd61d86 req-0767f3aa-5f3c-44ed-9859-566ac5992f0b service nova] Acquiring lock "refresh_cache-c29ad19e-91a5-4c3d-912b-bd7be4c42016" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.218450] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Acquiring lock "fd763e12-5b29-48f8-8256-cb6205e0d119" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.218836] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Lock "fd763e12-5b29-48f8-8256-cb6205e0d119" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.279988] env[62569]: DEBUG nova.network.neutron [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.303293] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249854, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.44846} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.303539] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 4b9e4f73-4682-4170-b71e-1d3d1d860577/4b9e4f73-4682-4170-b71e-1d3d1d860577.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 624.303791] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 624.303999] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4049b5de-0234-491f-841a-de430320ff74 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.309497] env[62569]: DEBUG nova.network.neutron [-] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.313165] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Waiting for the task: (returnval){ [ 624.313165] env[62569]: value = "task-1249855" [ 624.313165] env[62569]: _type = "Task" [ 624.313165] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.318552] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c9efccd-507d-4403-80ee-36302d3972c9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.328756] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1f21102-9afd-4600-8f8d-e779b0b7e6e7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.331828] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249855, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.361121] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-610d54d5-2781-4726-bec5-7292cd2970f6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.368295] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4169c9e4-6aa8-4ead-87b7-41e9a714d271 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.381477] env[62569]: DEBUG nova.compute.provider_tree [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.471033] env[62569]: DEBUG nova.network.neutron [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.783495] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Releasing lock "refresh_cache-c29ad19e-91a5-4c3d-912b-bd7be4c42016" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.784321] env[62569]: DEBUG nova.compute.manager [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 624.784567] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 624.784868] env[62569]: DEBUG oslo_concurrency.lockutils [req-1defd460-b47e-4224-af8c-042b4bd61d86 req-0767f3aa-5f3c-44ed-9859-566ac5992f0b service nova] Acquired lock "refresh_cache-c29ad19e-91a5-4c3d-912b-bd7be4c42016" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.785233] env[62569]: DEBUG nova.network.neutron [req-1defd460-b47e-4224-af8c-042b4bd61d86 req-0767f3aa-5f3c-44ed-9859-566ac5992f0b service nova] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Refreshing network info cache for port 78e8ea53-39d8-4665-a85e-497722b2d3c3 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 624.786116] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f2a96548-dc02-496a-b94b-b34a832671d9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.795279] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fd1f06a-b5ce-4b17-8188-0f95821ed157 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.817751] env[62569]: INFO nova.compute.manager [-] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Took 1.03 seconds to deallocate network for instance. [ 624.818910] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c29ad19e-91a5-4c3d-912b-bd7be4c42016 could not be found. [ 624.819168] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 624.819362] env[62569]: INFO nova.compute.manager [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Took 0.03 seconds to destroy the instance on the hypervisor. [ 624.819717] env[62569]: DEBUG oslo.service.loopingcall [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 624.824607] env[62569]: DEBUG nova.compute.manager [-] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 624.824712] env[62569]: DEBUG nova.network.neutron [-] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 624.831480] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249855, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062764} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.831584] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 624.832346] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24921cbe-678d-43d9-accb-739d23a6d254 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.851652] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Reconfiguring VM instance instance-00000019 to attach disk [datastore1] 4b9e4f73-4682-4170-b71e-1d3d1d860577/4b9e4f73-4682-4170-b71e-1d3d1d860577.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 624.852480] env[62569]: DEBUG nova.network.neutron [-] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.853705] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-78c3887b-b873-43b4-acf6-2f8cef7ab815 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.873203] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Waiting for the task: (returnval){ [ 624.873203] env[62569]: value = "task-1249856" [ 624.873203] env[62569]: _type = "Task" [ 624.873203] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.880637] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249856, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.884550] env[62569]: DEBUG nova.scheduler.client.report [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 624.973751] env[62569]: INFO nova.compute.manager [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 2a61b4cd-a15c-4064-9351-326dd777603b] Took 1.03 seconds to deallocate network for instance. [ 625.310252] env[62569]: DEBUG nova.network.neutron [req-1defd460-b47e-4224-af8c-042b4bd61d86 req-0767f3aa-5f3c-44ed-9859-566ac5992f0b service nova] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.368867] env[62569]: DEBUG nova.network.neutron [-] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.377799] env[62569]: INFO nova.compute.manager [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Took 0.56 seconds to detach 1 volumes for instance. [ 625.379961] env[62569]: DEBUG nova.network.neutron [req-1defd460-b47e-4224-af8c-042b4bd61d86 req-0767f3aa-5f3c-44ed-9859-566ac5992f0b service nova] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.382126] env[62569]: DEBUG nova.compute.claims [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 625.382383] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.387013] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249856, 'name': ReconfigVM_Task, 'duration_secs': 0.266564} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.387266] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Reconfigured VM instance instance-00000019 to attach disk [datastore1] 4b9e4f73-4682-4170-b71e-1d3d1d860577/4b9e4f73-4682-4170-b71e-1d3d1d860577.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 625.387831] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a9a019f0-17f7-4905-a7ca-bb4e94304489 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.389937] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.390367] env[62569]: DEBUG nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 625.393270] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.354s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.400515] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Waiting for the task: (returnval){ [ 625.400515] env[62569]: value = "task-1249857" [ 625.400515] env[62569]: _type = "Task" [ 625.400515] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.409091] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249857, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.873028] env[62569]: INFO nova.compute.manager [-] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Took 1.05 seconds to deallocate network for instance. [ 625.874841] env[62569]: DEBUG nova.compute.claims [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 625.875033] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.882573] env[62569]: DEBUG oslo_concurrency.lockutils [req-1defd460-b47e-4224-af8c-042b4bd61d86 req-0767f3aa-5f3c-44ed-9859-566ac5992f0b service nova] Releasing lock "refresh_cache-c29ad19e-91a5-4c3d-912b-bd7be4c42016" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.882796] env[62569]: DEBUG nova.compute.manager [req-1defd460-b47e-4224-af8c-042b4bd61d86 req-0767f3aa-5f3c-44ed-9859-566ac5992f0b service nova] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Received event network-vif-deleted-78e8ea53-39d8-4665-a85e-497722b2d3c3 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 625.897160] env[62569]: DEBUG nova.compute.utils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 625.902116] env[62569]: DEBUG nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 625.902116] env[62569]: DEBUG nova.network.neutron [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 625.913377] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249857, 'name': Rename_Task, 'duration_secs': 0.126067} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.913643] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 625.913910] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7dc8116a-7ef3-4ef9-863f-774870ee0dd8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.920662] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Waiting for the task: (returnval){ [ 625.920662] env[62569]: value = "task-1249858" [ 625.920662] env[62569]: _type = "Task" [ 625.920662] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.931462] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249858, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.946840] env[62569]: DEBUG nova.policy [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd72d243a83614002a7e300dc462c7b77', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58f52ff5374a4d5d9969aa2f48f01c8f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 626.001720] env[62569]: INFO nova.scheduler.client.report [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Deleted allocations for instance 2a61b4cd-a15c-4064-9351-326dd777603b [ 626.202459] env[62569]: DEBUG nova.network.neutron [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Successfully created port: 12465efc-17a2-432b-a4c4-ff14cf34a080 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 626.395544] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72baa3c9-17b7-4c43-b8d9-c9dad9a7d4f7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.404189] env[62569]: DEBUG nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 626.407485] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2fffd22-9566-4699-8b24-d4c099603b4c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.447138] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d298da2-da7c-49d8-aca2-5c2dbd615f5a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.457959] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b45c51e5-e3e0-476a-8e5a-13a8b9b74142 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.461796] env[62569]: DEBUG oslo_vmware.api [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249858, 'name': PowerOnVM_Task, 'duration_secs': 0.407511} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.462613] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 626.462778] env[62569]: DEBUG nova.compute.manager [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 626.463937] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89292ba7-834a-4550-b32e-66d99be19a51 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.474225] env[62569]: DEBUG nova.compute.provider_tree [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.511838] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a2ff640f-8c8c-4ad5-9562-07cccdd07ac4 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Lock "2a61b4cd-a15c-4064-9351-326dd777603b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 87.130s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.878284] env[62569]: DEBUG nova.compute.manager [req-23e5c967-db5f-4a4f-8ec7-32ae56dfb06a req-62845d2b-8fb0-4e0a-b95d-5ff4da867d36 service nova] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Received event network-changed-12465efc-17a2-432b-a4c4-ff14cf34a080 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 626.878284] env[62569]: DEBUG nova.compute.manager [req-23e5c967-db5f-4a4f-8ec7-32ae56dfb06a req-62845d2b-8fb0-4e0a-b95d-5ff4da867d36 service nova] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Refreshing instance network info cache due to event network-changed-12465efc-17a2-432b-a4c4-ff14cf34a080. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 626.878284] env[62569]: DEBUG oslo_concurrency.lockutils [req-23e5c967-db5f-4a4f-8ec7-32ae56dfb06a req-62845d2b-8fb0-4e0a-b95d-5ff4da867d36 service nova] Acquiring lock "refresh_cache-418c648a-06be-4e97-b138-45c65addf5cf" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.878284] env[62569]: DEBUG oslo_concurrency.lockutils [req-23e5c967-db5f-4a4f-8ec7-32ae56dfb06a req-62845d2b-8fb0-4e0a-b95d-5ff4da867d36 service nova] Acquired lock "refresh_cache-418c648a-06be-4e97-b138-45c65addf5cf" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.878284] env[62569]: DEBUG nova.network.neutron [req-23e5c967-db5f-4a4f-8ec7-32ae56dfb06a req-62845d2b-8fb0-4e0a-b95d-5ff4da867d36 service nova] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Refreshing network info cache for port 12465efc-17a2-432b-a4c4-ff14cf34a080 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 626.980936] env[62569]: DEBUG nova.scheduler.client.report [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 626.997618] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.014215] env[62569]: DEBUG nova.compute.manager [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 627.066612] env[62569]: ERROR nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 12465efc-17a2-432b-a4c4-ff14cf34a080, please check neutron logs for more information. [ 627.066612] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 627.066612] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 627.066612] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 627.066612] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 627.066612] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 627.066612] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 627.066612] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 627.066612] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.066612] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 627.066612] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.066612] env[62569]: ERROR nova.compute.manager raise self.value [ 627.066612] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 627.066612] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 627.066612] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.066612] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 627.067141] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.067141] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 627.067141] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 12465efc-17a2-432b-a4c4-ff14cf34a080, please check neutron logs for more information. [ 627.067141] env[62569]: ERROR nova.compute.manager [ 627.067141] env[62569]: Traceback (most recent call last): [ 627.067141] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 627.067141] env[62569]: listener.cb(fileno) [ 627.067141] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.067141] env[62569]: result = function(*args, **kwargs) [ 627.067141] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.067141] env[62569]: return func(*args, **kwargs) [ 627.067141] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 627.067141] env[62569]: raise e [ 627.067141] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 627.067141] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 627.067141] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 627.067141] env[62569]: created_port_ids = self._update_ports_for_instance( [ 627.067141] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 627.067141] env[62569]: with excutils.save_and_reraise_exception(): [ 627.067141] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.067141] env[62569]: self.force_reraise() [ 627.067141] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.067141] env[62569]: raise self.value [ 627.067141] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 627.067141] env[62569]: updated_port = self._update_port( [ 627.067141] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.067141] env[62569]: _ensure_no_port_binding_failure(port) [ 627.067141] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.067141] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 627.067944] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 12465efc-17a2-432b-a4c4-ff14cf34a080, please check neutron logs for more information. [ 627.067944] env[62569]: Removing descriptor: 17 [ 627.396735] env[62569]: DEBUG nova.network.neutron [req-23e5c967-db5f-4a4f-8ec7-32ae56dfb06a req-62845d2b-8fb0-4e0a-b95d-5ff4da867d36 service nova] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.419048] env[62569]: DEBUG nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 627.446846] env[62569]: DEBUG nova.virt.hardware [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 627.447250] env[62569]: DEBUG nova.virt.hardware [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 627.447513] env[62569]: DEBUG nova.virt.hardware [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 627.447835] env[62569]: DEBUG nova.virt.hardware [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 627.448134] env[62569]: DEBUG nova.virt.hardware [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 627.448404] env[62569]: DEBUG nova.virt.hardware [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 627.448744] env[62569]: DEBUG nova.virt.hardware [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 627.449023] env[62569]: DEBUG nova.virt.hardware [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 627.449318] env[62569]: DEBUG nova.virt.hardware [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 627.449607] env[62569]: DEBUG nova.virt.hardware [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 627.449922] env[62569]: DEBUG nova.virt.hardware [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 627.451237] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-801fa339-e2e7-4daa-ab1a-783f40c36873 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.459785] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa86a047-1806-46a7-b13c-d41b2e23f1ca {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.473317] env[62569]: ERROR nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 12465efc-17a2-432b-a4c4-ff14cf34a080, please check neutron logs for more information. [ 627.473317] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Traceback (most recent call last): [ 627.473317] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 627.473317] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] yield resources [ 627.473317] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 627.473317] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] self.driver.spawn(context, instance, image_meta, [ 627.473317] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 627.473317] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 627.473317] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 627.473317] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] vm_ref = self.build_virtual_machine(instance, [ 627.473317] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 627.473680] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] vif_infos = vmwarevif.get_vif_info(self._session, [ 627.473680] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 627.473680] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] for vif in network_info: [ 627.473680] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 627.473680] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] return self._sync_wrapper(fn, *args, **kwargs) [ 627.473680] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 627.473680] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] self.wait() [ 627.473680] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 627.473680] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] self[:] = self._gt.wait() [ 627.473680] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 627.473680] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] return self._exit_event.wait() [ 627.473680] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 627.473680] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] current.throw(*self._exc) [ 627.474051] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.474051] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] result = function(*args, **kwargs) [ 627.474051] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.474051] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] return func(*args, **kwargs) [ 627.474051] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 627.474051] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] raise e [ 627.474051] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 627.474051] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] nwinfo = self.network_api.allocate_for_instance( [ 627.474051] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 627.474051] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] created_port_ids = self._update_ports_for_instance( [ 627.474051] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 627.474051] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] with excutils.save_and_reraise_exception(): [ 627.474051] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.474408] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] self.force_reraise() [ 627.474408] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.474408] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] raise self.value [ 627.474408] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 627.474408] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] updated_port = self._update_port( [ 627.474408] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.474408] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] _ensure_no_port_binding_failure(port) [ 627.474408] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.474408] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] raise exception.PortBindingFailed(port_id=port['id']) [ 627.474408] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] nova.exception.PortBindingFailed: Binding failed for port 12465efc-17a2-432b-a4c4-ff14cf34a080, please check neutron logs for more information. [ 627.474408] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] [ 627.474408] env[62569]: INFO nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Terminating instance [ 627.475661] env[62569]: DEBUG nova.network.neutron [req-23e5c967-db5f-4a4f-8ec7-32ae56dfb06a req-62845d2b-8fb0-4e0a-b95d-5ff4da867d36 service nova] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.492074] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.099s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.492548] env[62569]: ERROR nova.compute.manager [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4d5e63a8-f88c-4b57-93a4-cc764ca6a405, please check neutron logs for more information. [ 627.492548] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Traceback (most recent call last): [ 627.492548] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 627.492548] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] self.driver.spawn(context, instance, image_meta, [ 627.492548] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 627.492548] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] self._vmops.spawn(context, instance, image_meta, injected_files, [ 627.492548] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 627.492548] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] vm_ref = self.build_virtual_machine(instance, [ 627.492548] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 627.492548] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] vif_infos = vmwarevif.get_vif_info(self._session, [ 627.492548] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 627.493015] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] for vif in network_info: [ 627.493015] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 627.493015] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] return self._sync_wrapper(fn, *args, **kwargs) [ 627.493015] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 627.493015] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] self.wait() [ 627.493015] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 627.493015] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] self[:] = self._gt.wait() [ 627.493015] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 627.493015] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] return self._exit_event.wait() [ 627.493015] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 627.493015] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] current.throw(*self._exc) [ 627.493015] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.493015] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] result = function(*args, **kwargs) [ 627.493370] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.493370] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] return func(*args, **kwargs) [ 627.493370] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 627.493370] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] raise e [ 627.493370] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 627.493370] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] nwinfo = self.network_api.allocate_for_instance( [ 627.493370] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 627.493370] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] created_port_ids = self._update_ports_for_instance( [ 627.493370] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 627.493370] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] with excutils.save_and_reraise_exception(): [ 627.493370] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.493370] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] self.force_reraise() [ 627.493370] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.493734] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] raise self.value [ 627.493734] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 627.493734] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] updated_port = self._update_port( [ 627.493734] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.493734] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] _ensure_no_port_binding_failure(port) [ 627.493734] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.493734] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] raise exception.PortBindingFailed(port_id=port['id']) [ 627.493734] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] nova.exception.PortBindingFailed: Binding failed for port 4d5e63a8-f88c-4b57-93a4-cc764ca6a405, please check neutron logs for more information. [ 627.493734] env[62569]: ERROR nova.compute.manager [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] [ 627.493734] env[62569]: DEBUG nova.compute.utils [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Binding failed for port 4d5e63a8-f88c-4b57-93a4-cc764ca6a405, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 627.494260] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.900s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.495668] env[62569]: INFO nova.compute.claims [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 627.498426] env[62569]: DEBUG nova.compute.manager [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Build of instance 76bddf93-ca4f-4987-a446-b369801ebd93 was re-scheduled: Binding failed for port 4d5e63a8-f88c-4b57-93a4-cc764ca6a405, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 627.498841] env[62569]: DEBUG nova.compute.manager [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 627.499076] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Acquiring lock "refresh_cache-76bddf93-ca4f-4987-a446-b369801ebd93" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.499223] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Acquired lock "refresh_cache-76bddf93-ca4f-4987-a446-b369801ebd93" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.499379] env[62569]: DEBUG nova.network.neutron [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 627.538600] env[62569]: DEBUG oslo_concurrency.lockutils [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.697056] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Acquiring lock "4b9e4f73-4682-4170-b71e-1d3d1d860577" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.697199] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Lock "4b9e4f73-4682-4170-b71e-1d3d1d860577" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.697752] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Acquiring lock "4b9e4f73-4682-4170-b71e-1d3d1d860577-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.697983] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Lock "4b9e4f73-4682-4170-b71e-1d3d1d860577-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.698187] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Lock "4b9e4f73-4682-4170-b71e-1d3d1d860577-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.701118] env[62569]: INFO nova.compute.manager [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Terminating instance [ 627.978435] env[62569]: DEBUG oslo_concurrency.lockutils [req-23e5c967-db5f-4a4f-8ec7-32ae56dfb06a req-62845d2b-8fb0-4e0a-b95d-5ff4da867d36 service nova] Releasing lock "refresh_cache-418c648a-06be-4e97-b138-45c65addf5cf" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.979032] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "refresh_cache-418c648a-06be-4e97-b138-45c65addf5cf" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.979202] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquired lock "refresh_cache-418c648a-06be-4e97-b138-45c65addf5cf" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.979379] env[62569]: DEBUG nova.network.neutron [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 628.016678] env[62569]: DEBUG nova.network.neutron [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.071661] env[62569]: DEBUG nova.network.neutron [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.205930] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Acquiring lock "refresh_cache-4b9e4f73-4682-4170-b71e-1d3d1d860577" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.205930] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Acquired lock "refresh_cache-4b9e4f73-4682-4170-b71e-1d3d1d860577" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.205930] env[62569]: DEBUG nova.network.neutron [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 628.497681] env[62569]: DEBUG nova.network.neutron [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.572296] env[62569]: DEBUG nova.network.neutron [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.573825] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Releasing lock "refresh_cache-76bddf93-ca4f-4987-a446-b369801ebd93" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.574065] env[62569]: DEBUG nova.compute.manager [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 628.574234] env[62569]: DEBUG nova.compute.manager [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 628.574399] env[62569]: DEBUG nova.network.neutron [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 628.590109] env[62569]: DEBUG nova.network.neutron [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.724983] env[62569]: DEBUG nova.network.neutron [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.775268] env[62569]: DEBUG nova.network.neutron [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.900897] env[62569]: DEBUG nova.compute.manager [req-980253d9-5cda-4bc9-a394-05eff5c81abf req-e556ca58-2868-4f65-a4bb-a09037d2e883 service nova] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Received event network-vif-deleted-12465efc-17a2-432b-a4c4-ff14cf34a080 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 628.968646] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d67b4d4-0a33-4ac8-baa1-2732259fa3fb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.978949] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8ef4d1d-1940-4b57-9368-ac9f440deab5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.010053] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2571f367-6d42-4061-a18f-2c7283fbc40f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.018020] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dae17c1-a57e-45d8-b8a6-eb913d7f0853 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.031055] env[62569]: DEBUG nova.compute.provider_tree [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.077827] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Releasing lock "refresh_cache-418c648a-06be-4e97-b138-45c65addf5cf" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.077827] env[62569]: DEBUG nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 629.078782] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 629.078782] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e6809a10-379e-4f8f-88ce-59c041e623e4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.087176] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38eb0446-5ba1-4ecc-9111-cf52d421e31a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.097279] env[62569]: DEBUG nova.network.neutron [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.110309] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 418c648a-06be-4e97-b138-45c65addf5cf could not be found. [ 629.110565] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 629.110734] env[62569]: INFO nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Took 0.03 seconds to destroy the instance on the hypervisor. [ 629.110988] env[62569]: DEBUG oslo.service.loopingcall [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 629.111848] env[62569]: DEBUG nova.compute.manager [-] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 629.111993] env[62569]: DEBUG nova.network.neutron [-] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 629.130342] env[62569]: DEBUG nova.network.neutron [-] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.278230] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Releasing lock "refresh_cache-4b9e4f73-4682-4170-b71e-1d3d1d860577" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.278717] env[62569]: DEBUG nova.compute.manager [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 629.278862] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 629.279865] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bbe9b41-7f31-4e3a-9a8e-d8e636fe9fed {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.288640] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 629.288835] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-746c7b9f-533b-48b4-9381-8bf4762846aa {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.295685] env[62569]: DEBUG oslo_vmware.api [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Waiting for the task: (returnval){ [ 629.295685] env[62569]: value = "task-1249859" [ 629.295685] env[62569]: _type = "Task" [ 629.295685] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.304928] env[62569]: DEBUG oslo_vmware.api [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249859, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.534528] env[62569]: DEBUG nova.scheduler.client.report [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 629.601615] env[62569]: INFO nova.compute.manager [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] [instance: 76bddf93-ca4f-4987-a446-b369801ebd93] Took 1.03 seconds to deallocate network for instance. [ 629.632407] env[62569]: DEBUG nova.network.neutron [-] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.806080] env[62569]: DEBUG oslo_vmware.api [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249859, 'name': PowerOffVM_Task, 'duration_secs': 0.119156} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.806080] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 629.806080] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 629.806080] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8929abc7-4ca2-4f2f-81d6-3181b846e62e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.831092] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 629.831262] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 629.831447] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Deleting the datastore file [datastore1] 4b9e4f73-4682-4170-b71e-1d3d1d860577 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 629.831705] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2e80eea4-2a49-4834-ab05-1236cc368b9d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.837775] env[62569]: DEBUG oslo_vmware.api [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Waiting for the task: (returnval){ [ 629.837775] env[62569]: value = "task-1249861" [ 629.837775] env[62569]: _type = "Task" [ 629.837775] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.845061] env[62569]: DEBUG oslo_vmware.api [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249861, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.042681] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.548s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.043240] env[62569]: DEBUG nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 630.045856] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.187s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.046045] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.046201] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62569) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 630.046515] env[62569]: DEBUG oslo_concurrency.lockutils [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.963s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.048861] env[62569]: INFO nova.compute.claims [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 630.053433] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39cd8686-bbd5-4efc-9633-cce014ac9f53 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.064365] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18eefea-2e38-4c83-b34b-a93ef1423f41 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.078178] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89463a0f-1bc4-4b30-a809-02512be36a2d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.085421] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-420d987f-2451-4977-ae67-6d390fa87e6b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.117680] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181537MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=62569) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 630.117835] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.135432] env[62569]: INFO nova.compute.manager [-] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Took 1.02 seconds to deallocate network for instance. [ 630.137793] env[62569]: DEBUG nova.compute.claims [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 630.137973] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.348655] env[62569]: DEBUG oslo_vmware.api [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Task: {'id': task-1249861, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.08835} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.348841] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 630.349036] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 630.349218] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 630.349390] env[62569]: INFO nova.compute.manager [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Took 1.07 seconds to destroy the instance on the hypervisor. [ 630.349677] env[62569]: DEBUG oslo.service.loopingcall [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 630.349865] env[62569]: DEBUG nova.compute.manager [-] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 630.349962] env[62569]: DEBUG nova.network.neutron [-] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 630.365441] env[62569]: DEBUG nova.network.neutron [-] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.554325] env[62569]: DEBUG nova.compute.utils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 630.555887] env[62569]: DEBUG nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 630.555965] env[62569]: DEBUG nova.network.neutron [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 630.598972] env[62569]: DEBUG nova.policy [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd72d243a83614002a7e300dc462c7b77', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58f52ff5374a4d5d9969aa2f48f01c8f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 630.641570] env[62569]: INFO nova.scheduler.client.report [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Deleted allocations for instance 76bddf93-ca4f-4987-a446-b369801ebd93 [ 630.857008] env[62569]: DEBUG nova.network.neutron [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Successfully created port: bc3b6680-2a28-49ee-8a81-8a88f2dbe09f {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 630.867446] env[62569]: DEBUG nova.network.neutron [-] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.060081] env[62569]: DEBUG nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 631.151859] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12789b81-91de-441f-b88f-dae468f1b5f6 tempest-ListImageFiltersTestJSON-1498332022 tempest-ListImageFiltersTestJSON-1498332022-project-member] Lock "76bddf93-ca4f-4987-a446-b369801ebd93" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.204s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.369894] env[62569]: INFO nova.compute.manager [-] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Took 1.02 seconds to deallocate network for instance. [ 631.444957] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-265465be-ef8f-4b4f-aa0d-ad409611e1e7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.452854] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edec2895-f48c-447d-8674-fe4467e82bed {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.483213] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd7f97b3-f08d-4eee-bbc3-272c55cbccbc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.490902] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-196ede4f-3a3b-4315-be71-039f2d193e0e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.503948] env[62569]: DEBUG nova.compute.provider_tree [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 631.537346] env[62569]: DEBUG nova.compute.manager [req-7edbed99-3e52-4e61-a9a7-03b21357afc6 req-f149c138-a02f-4220-bb3f-b8f81b9e5f89 service nova] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Received event network-changed-bc3b6680-2a28-49ee-8a81-8a88f2dbe09f {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 631.537565] env[62569]: DEBUG nova.compute.manager [req-7edbed99-3e52-4e61-a9a7-03b21357afc6 req-f149c138-a02f-4220-bb3f-b8f81b9e5f89 service nova] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Refreshing instance network info cache due to event network-changed-bc3b6680-2a28-49ee-8a81-8a88f2dbe09f. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 631.537821] env[62569]: DEBUG oslo_concurrency.lockutils [req-7edbed99-3e52-4e61-a9a7-03b21357afc6 req-f149c138-a02f-4220-bb3f-b8f81b9e5f89 service nova] Acquiring lock "refresh_cache-92666c7b-7a1d-4d49-931f-e84cd2486871" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.537984] env[62569]: DEBUG oslo_concurrency.lockutils [req-7edbed99-3e52-4e61-a9a7-03b21357afc6 req-f149c138-a02f-4220-bb3f-b8f81b9e5f89 service nova] Acquired lock "refresh_cache-92666c7b-7a1d-4d49-931f-e84cd2486871" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.538164] env[62569]: DEBUG nova.network.neutron [req-7edbed99-3e52-4e61-a9a7-03b21357afc6 req-f149c138-a02f-4220-bb3f-b8f81b9e5f89 service nova] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Refreshing network info cache for port bc3b6680-2a28-49ee-8a81-8a88f2dbe09f {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 631.655565] env[62569]: DEBUG nova.compute.manager [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 631.699437] env[62569]: ERROR nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bc3b6680-2a28-49ee-8a81-8a88f2dbe09f, please check neutron logs for more information. [ 631.699437] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 631.699437] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 631.699437] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 631.699437] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.699437] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 631.699437] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.699437] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 631.699437] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.699437] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 631.699437] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.699437] env[62569]: ERROR nova.compute.manager raise self.value [ 631.699437] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.699437] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 631.699437] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.699437] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 631.700105] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.700105] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 631.700105] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bc3b6680-2a28-49ee-8a81-8a88f2dbe09f, please check neutron logs for more information. [ 631.700105] env[62569]: ERROR nova.compute.manager [ 631.700105] env[62569]: Traceback (most recent call last): [ 631.700105] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 631.700105] env[62569]: listener.cb(fileno) [ 631.700105] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.700105] env[62569]: result = function(*args, **kwargs) [ 631.700105] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 631.700105] env[62569]: return func(*args, **kwargs) [ 631.700105] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 631.700105] env[62569]: raise e [ 631.700105] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 631.700105] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 631.700105] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.700105] env[62569]: created_port_ids = self._update_ports_for_instance( [ 631.700105] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.700105] env[62569]: with excutils.save_and_reraise_exception(): [ 631.700105] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.700105] env[62569]: self.force_reraise() [ 631.700105] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.700105] env[62569]: raise self.value [ 631.700105] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.700105] env[62569]: updated_port = self._update_port( [ 631.700105] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.700105] env[62569]: _ensure_no_port_binding_failure(port) [ 631.700105] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.700105] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 631.700821] env[62569]: nova.exception.PortBindingFailed: Binding failed for port bc3b6680-2a28-49ee-8a81-8a88f2dbe09f, please check neutron logs for more information. [ 631.700821] env[62569]: Removing descriptor: 17 [ 631.877768] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.010566] env[62569]: DEBUG nova.scheduler.client.report [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 632.057048] env[62569]: DEBUG nova.network.neutron [req-7edbed99-3e52-4e61-a9a7-03b21357afc6 req-f149c138-a02f-4220-bb3f-b8f81b9e5f89 service nova] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.075121] env[62569]: DEBUG nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 632.099490] env[62569]: DEBUG nova.virt.hardware [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 632.099753] env[62569]: DEBUG nova.virt.hardware [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 632.099920] env[62569]: DEBUG nova.virt.hardware [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 632.100124] env[62569]: DEBUG nova.virt.hardware [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 632.100274] env[62569]: DEBUG nova.virt.hardware [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 632.100420] env[62569]: DEBUG nova.virt.hardware [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 632.100646] env[62569]: DEBUG nova.virt.hardware [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 632.100829] env[62569]: DEBUG nova.virt.hardware [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 632.101209] env[62569]: DEBUG nova.virt.hardware [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 632.101209] env[62569]: DEBUG nova.virt.hardware [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 632.101368] env[62569]: DEBUG nova.virt.hardware [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 632.102229] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-889dba50-ecf1-4073-a9a9-5528a7626b85 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.112657] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e1663b-e293-4428-bbcb-ebeecdffd29a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.126292] env[62569]: ERROR nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bc3b6680-2a28-49ee-8a81-8a88f2dbe09f, please check neutron logs for more information. [ 632.126292] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Traceback (most recent call last): [ 632.126292] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 632.126292] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] yield resources [ 632.126292] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 632.126292] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] self.driver.spawn(context, instance, image_meta, [ 632.126292] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 632.126292] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] self._vmops.spawn(context, instance, image_meta, injected_files, [ 632.126292] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 632.126292] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] vm_ref = self.build_virtual_machine(instance, [ 632.126292] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 632.126892] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] vif_infos = vmwarevif.get_vif_info(self._session, [ 632.126892] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 632.126892] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] for vif in network_info: [ 632.126892] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 632.126892] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] return self._sync_wrapper(fn, *args, **kwargs) [ 632.126892] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 632.126892] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] self.wait() [ 632.126892] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 632.126892] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] self[:] = self._gt.wait() [ 632.126892] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 632.126892] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] return self._exit_event.wait() [ 632.126892] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 632.126892] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] current.throw(*self._exc) [ 632.127304] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.127304] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] result = function(*args, **kwargs) [ 632.127304] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.127304] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] return func(*args, **kwargs) [ 632.127304] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 632.127304] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] raise e [ 632.127304] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 632.127304] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] nwinfo = self.network_api.allocate_for_instance( [ 632.127304] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.127304] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] created_port_ids = self._update_ports_for_instance( [ 632.127304] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.127304] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] with excutils.save_and_reraise_exception(): [ 632.127304] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.127659] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] self.force_reraise() [ 632.127659] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.127659] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] raise self.value [ 632.127659] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.127659] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] updated_port = self._update_port( [ 632.127659] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.127659] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] _ensure_no_port_binding_failure(port) [ 632.127659] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.127659] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] raise exception.PortBindingFailed(port_id=port['id']) [ 632.127659] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] nova.exception.PortBindingFailed: Binding failed for port bc3b6680-2a28-49ee-8a81-8a88f2dbe09f, please check neutron logs for more information. [ 632.127659] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] [ 632.127659] env[62569]: INFO nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Terminating instance [ 632.143594] env[62569]: DEBUG nova.network.neutron [req-7edbed99-3e52-4e61-a9a7-03b21357afc6 req-f149c138-a02f-4220-bb3f-b8f81b9e5f89 service nova] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.179648] env[62569]: DEBUG oslo_concurrency.lockutils [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.516027] env[62569]: DEBUG oslo_concurrency.lockutils [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.469s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.516662] env[62569]: DEBUG nova.compute.manager [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 632.519321] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.875s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.632206] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "refresh_cache-92666c7b-7a1d-4d49-931f-e84cd2486871" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.646517] env[62569]: DEBUG oslo_concurrency.lockutils [req-7edbed99-3e52-4e61-a9a7-03b21357afc6 req-f149c138-a02f-4220-bb3f-b8f81b9e5f89 service nova] Releasing lock "refresh_cache-92666c7b-7a1d-4d49-931f-e84cd2486871" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.646517] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquired lock "refresh_cache-92666c7b-7a1d-4d49-931f-e84cd2486871" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.646517] env[62569]: DEBUG nova.network.neutron [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 633.028494] env[62569]: DEBUG nova.compute.utils [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 633.044435] env[62569]: DEBUG nova.compute.manager [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 633.044618] env[62569]: DEBUG nova.network.neutron [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 633.128273] env[62569]: DEBUG nova.policy [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd1fb952884374432b2ce857f87f101f1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c5fd913659414266bc6092b849e16da9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 633.167790] env[62569]: DEBUG nova.network.neutron [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.291617] env[62569]: DEBUG nova.network.neutron [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.509517] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d83a3bfd-640a-4818-908e-54cbde7d7464 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.517161] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f5e8a0f-e8f5-49e2-a5b8-d601f7840cd0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.549630] env[62569]: DEBUG nova.compute.manager [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 633.552771] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0806da2d-367e-4a51-b582-1bee638fcc10 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.563914] env[62569]: DEBUG nova.network.neutron [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Successfully created port: 529ffd30-aece-4729-b8da-d28c3ef597d9 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 633.570140] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d98eb8-acc8-456f-8977-ff7aba058a87 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.574919] env[62569]: DEBUG nova.compute.manager [req-6219795b-2919-4c7e-b78f-308a6a982c13 req-0cc702b1-b60b-45f6-a341-8ab29fd655c4 service nova] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Received event network-vif-deleted-bc3b6680-2a28-49ee-8a81-8a88f2dbe09f {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 633.586708] env[62569]: DEBUG nova.compute.provider_tree [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 633.797043] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Releasing lock "refresh_cache-92666c7b-7a1d-4d49-931f-e84cd2486871" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.797491] env[62569]: DEBUG nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 633.797707] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 633.798140] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4d3c5da7-801f-4d52-bb7d-21e2e871998b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.814207] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-969c9443-b6d6-416e-ac03-2f472e06e4cd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.841787] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 92666c7b-7a1d-4d49-931f-e84cd2486871 could not be found. [ 633.842380] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 633.842417] env[62569]: INFO nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Took 0.04 seconds to destroy the instance on the hypervisor. [ 633.844055] env[62569]: DEBUG oslo.service.loopingcall [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 633.844711] env[62569]: DEBUG nova.compute.manager [-] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 633.844993] env[62569]: DEBUG nova.network.neutron [-] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 633.868229] env[62569]: DEBUG nova.network.neutron [-] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.092165] env[62569]: DEBUG nova.scheduler.client.report [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 634.370863] env[62569]: DEBUG nova.network.neutron [-] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.468249] env[62569]: ERROR nova.compute.manager [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 529ffd30-aece-4729-b8da-d28c3ef597d9, please check neutron logs for more information. [ 634.468249] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 634.468249] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 634.468249] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 634.468249] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 634.468249] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 634.468249] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 634.468249] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 634.468249] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.468249] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 634.468249] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.468249] env[62569]: ERROR nova.compute.manager raise self.value [ 634.468249] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 634.468249] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 634.468249] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.468249] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 634.468749] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.468749] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 634.468749] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 529ffd30-aece-4729-b8da-d28c3ef597d9, please check neutron logs for more information. [ 634.468749] env[62569]: ERROR nova.compute.manager [ 634.468749] env[62569]: Traceback (most recent call last): [ 634.468749] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 634.468749] env[62569]: listener.cb(fileno) [ 634.468749] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.468749] env[62569]: result = function(*args, **kwargs) [ 634.468749] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 634.468749] env[62569]: return func(*args, **kwargs) [ 634.468749] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 634.468749] env[62569]: raise e [ 634.468749] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 634.468749] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 634.468749] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 634.468749] env[62569]: created_port_ids = self._update_ports_for_instance( [ 634.468749] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 634.468749] env[62569]: with excutils.save_and_reraise_exception(): [ 634.468749] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.468749] env[62569]: self.force_reraise() [ 634.468749] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.468749] env[62569]: raise self.value [ 634.468749] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 634.468749] env[62569]: updated_port = self._update_port( [ 634.468749] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.468749] env[62569]: _ensure_no_port_binding_failure(port) [ 634.468749] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.468749] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 634.469562] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 529ffd30-aece-4729-b8da-d28c3ef597d9, please check neutron logs for more information. [ 634.469562] env[62569]: Removing descriptor: 17 [ 634.565394] env[62569]: DEBUG nova.compute.manager [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 634.591601] env[62569]: DEBUG nova.virt.hardware [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 634.591840] env[62569]: DEBUG nova.virt.hardware [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 634.592018] env[62569]: DEBUG nova.virt.hardware [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 634.592223] env[62569]: DEBUG nova.virt.hardware [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 634.592365] env[62569]: DEBUG nova.virt.hardware [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 634.592507] env[62569]: DEBUG nova.virt.hardware [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 634.592716] env[62569]: DEBUG nova.virt.hardware [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 634.592870] env[62569]: DEBUG nova.virt.hardware [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 634.593047] env[62569]: DEBUG nova.virt.hardware [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 634.593212] env[62569]: DEBUG nova.virt.hardware [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 634.593397] env[62569]: DEBUG nova.virt.hardware [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 634.594345] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee6cccff-a7c1-42cd-a686-5a554869ec73 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.601776] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.083s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.602385] env[62569]: ERROR nova.compute.manager [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b46ffb38-0127-4d26-812f-fb3830415886, please check neutron logs for more information. [ 634.602385] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Traceback (most recent call last): [ 634.602385] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 634.602385] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] self.driver.spawn(context, instance, image_meta, [ 634.602385] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 634.602385] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 634.602385] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 634.602385] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] vm_ref = self.build_virtual_machine(instance, [ 634.602385] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 634.602385] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] vif_infos = vmwarevif.get_vif_info(self._session, [ 634.602385] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 634.602714] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] for vif in network_info: [ 634.602714] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 634.602714] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] return self._sync_wrapper(fn, *args, **kwargs) [ 634.602714] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 634.602714] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] self.wait() [ 634.602714] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 634.602714] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] self[:] = self._gt.wait() [ 634.602714] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 634.602714] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] return self._exit_event.wait() [ 634.602714] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 634.602714] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] current.throw(*self._exc) [ 634.602714] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.602714] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] result = function(*args, **kwargs) [ 634.603291] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 634.603291] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] return func(*args, **kwargs) [ 634.603291] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 634.603291] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] raise e [ 634.603291] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 634.603291] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] nwinfo = self.network_api.allocate_for_instance( [ 634.603291] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 634.603291] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] created_port_ids = self._update_ports_for_instance( [ 634.603291] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 634.603291] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] with excutils.save_and_reraise_exception(): [ 634.603291] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.603291] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] self.force_reraise() [ 634.603291] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.603720] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] raise self.value [ 634.603720] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 634.603720] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] updated_port = self._update_port( [ 634.603720] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.603720] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] _ensure_no_port_binding_failure(port) [ 634.603720] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.603720] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] raise exception.PortBindingFailed(port_id=port['id']) [ 634.603720] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] nova.exception.PortBindingFailed: Binding failed for port b46ffb38-0127-4d26-812f-fb3830415886, please check neutron logs for more information. [ 634.603720] env[62569]: ERROR nova.compute.manager [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] [ 634.603720] env[62569]: DEBUG nova.compute.utils [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Binding failed for port b46ffb38-0127-4d26-812f-fb3830415886, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 634.604314] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.335s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.605645] env[62569]: INFO nova.compute.claims [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 634.609301] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6736cfc2-e9b2-4972-912d-39cfdb32c7ad {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.613381] env[62569]: DEBUG nova.compute.manager [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Build of instance 898f69fe-737a-44cb-9c43-965f1f8685f7 was re-scheduled: Binding failed for port b46ffb38-0127-4d26-812f-fb3830415886, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 634.613824] env[62569]: DEBUG nova.compute.manager [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 634.614071] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "refresh_cache-898f69fe-737a-44cb-9c43-965f1f8685f7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.614234] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquired lock "refresh_cache-898f69fe-737a-44cb-9c43-965f1f8685f7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.614393] env[62569]: DEBUG nova.network.neutron [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 634.627467] env[62569]: ERROR nova.compute.manager [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 529ffd30-aece-4729-b8da-d28c3ef597d9, please check neutron logs for more information. [ 634.627467] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Traceback (most recent call last): [ 634.627467] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 634.627467] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] yield resources [ 634.627467] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 634.627467] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] self.driver.spawn(context, instance, image_meta, [ 634.627467] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 634.627467] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 634.627467] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 634.627467] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] vm_ref = self.build_virtual_machine(instance, [ 634.627467] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 634.627912] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] vif_infos = vmwarevif.get_vif_info(self._session, [ 634.627912] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 634.627912] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] for vif in network_info: [ 634.627912] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 634.627912] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] return self._sync_wrapper(fn, *args, **kwargs) [ 634.627912] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 634.627912] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] self.wait() [ 634.627912] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 634.627912] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] self[:] = self._gt.wait() [ 634.627912] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 634.627912] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] return self._exit_event.wait() [ 634.627912] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 634.627912] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] current.throw(*self._exc) [ 634.628298] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.628298] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] result = function(*args, **kwargs) [ 634.628298] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 634.628298] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] return func(*args, **kwargs) [ 634.628298] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 634.628298] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] raise e [ 634.628298] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 634.628298] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] nwinfo = self.network_api.allocate_for_instance( [ 634.628298] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 634.628298] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] created_port_ids = self._update_ports_for_instance( [ 634.628298] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 634.628298] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] with excutils.save_and_reraise_exception(): [ 634.628298] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.628661] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] self.force_reraise() [ 634.628661] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.628661] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] raise self.value [ 634.628661] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 634.628661] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] updated_port = self._update_port( [ 634.628661] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.628661] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] _ensure_no_port_binding_failure(port) [ 634.628661] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.628661] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] raise exception.PortBindingFailed(port_id=port['id']) [ 634.628661] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] nova.exception.PortBindingFailed: Binding failed for port 529ffd30-aece-4729-b8da-d28c3ef597d9, please check neutron logs for more information. [ 634.628661] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] [ 634.628661] env[62569]: INFO nova.compute.manager [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Terminating instance [ 634.873526] env[62569]: INFO nova.compute.manager [-] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Took 1.03 seconds to deallocate network for instance. [ 634.876996] env[62569]: DEBUG nova.compute.claims [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 634.877210] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.132189] env[62569]: DEBUG oslo_concurrency.lockutils [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Acquiring lock "refresh_cache-21cde60b-2764-4ff7-ab5e-a17fb672351d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.132690] env[62569]: DEBUG oslo_concurrency.lockutils [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Acquired lock "refresh_cache-21cde60b-2764-4ff7-ab5e-a17fb672351d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.132690] env[62569]: DEBUG nova.network.neutron [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 635.143327] env[62569]: DEBUG nova.network.neutron [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 635.226306] env[62569]: DEBUG nova.network.neutron [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.593303] env[62569]: DEBUG nova.compute.manager [req-830eaa62-9199-4c97-9d1f-5e0149d6bf06 req-a9d7c269-ec98-4e9a-812d-f4d0f4ac6deb service nova] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Received event network-changed-529ffd30-aece-4729-b8da-d28c3ef597d9 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 635.593495] env[62569]: DEBUG nova.compute.manager [req-830eaa62-9199-4c97-9d1f-5e0149d6bf06 req-a9d7c269-ec98-4e9a-812d-f4d0f4ac6deb service nova] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Refreshing instance network info cache due to event network-changed-529ffd30-aece-4729-b8da-d28c3ef597d9. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 635.593738] env[62569]: DEBUG oslo_concurrency.lockutils [req-830eaa62-9199-4c97-9d1f-5e0149d6bf06 req-a9d7c269-ec98-4e9a-812d-f4d0f4ac6deb service nova] Acquiring lock "refresh_cache-21cde60b-2764-4ff7-ab5e-a17fb672351d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.663542] env[62569]: DEBUG nova.network.neutron [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 635.730980] env[62569]: DEBUG nova.network.neutron [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.736482] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Releasing lock "refresh_cache-898f69fe-737a-44cb-9c43-965f1f8685f7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.736682] env[62569]: DEBUG nova.compute.manager [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 635.736852] env[62569]: DEBUG nova.compute.manager [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 635.737060] env[62569]: DEBUG nova.network.neutron [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 635.755544] env[62569]: DEBUG nova.network.neutron [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 636.128393] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24166067-45de-4d17-aa92-5f7ddd6cb436 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.139467] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-106039fc-a96f-44dc-b583-a3430fd3a968 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.169831] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb2bab8-8b72-467e-bce4-1c0462d11093 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.177251] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-989574a3-8628-437b-95d4-723574ed3a0a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.190790] env[62569]: DEBUG nova.compute.provider_tree [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 636.233826] env[62569]: DEBUG oslo_concurrency.lockutils [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Releasing lock "refresh_cache-21cde60b-2764-4ff7-ab5e-a17fb672351d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.234362] env[62569]: DEBUG nova.compute.manager [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 636.234560] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 636.234865] env[62569]: DEBUG oslo_concurrency.lockutils [req-830eaa62-9199-4c97-9d1f-5e0149d6bf06 req-a9d7c269-ec98-4e9a-812d-f4d0f4ac6deb service nova] Acquired lock "refresh_cache-21cde60b-2764-4ff7-ab5e-a17fb672351d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.235047] env[62569]: DEBUG nova.network.neutron [req-830eaa62-9199-4c97-9d1f-5e0149d6bf06 req-a9d7c269-ec98-4e9a-812d-f4d0f4ac6deb service nova] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Refreshing network info cache for port 529ffd30-aece-4729-b8da-d28c3ef597d9 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 636.236388] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2c794358-0287-4f04-81d1-df634d7d1ad3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.246475] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2fa8f8e-f9e3-4561-95ac-14a94c15f61f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.259909] env[62569]: DEBUG nova.network.neutron [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.269474] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 21cde60b-2764-4ff7-ab5e-a17fb672351d could not be found. [ 636.269658] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 636.269859] env[62569]: INFO nova.compute.manager [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 636.270118] env[62569]: DEBUG oslo.service.loopingcall [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 636.270339] env[62569]: DEBUG nova.compute.manager [-] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 636.270419] env[62569]: DEBUG nova.network.neutron [-] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 636.288138] env[62569]: DEBUG nova.network.neutron [-] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 636.694725] env[62569]: DEBUG nova.scheduler.client.report [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 636.758481] env[62569]: DEBUG nova.network.neutron [req-830eaa62-9199-4c97-9d1f-5e0149d6bf06 req-a9d7c269-ec98-4e9a-812d-f4d0f4ac6deb service nova] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 636.762941] env[62569]: INFO nova.compute.manager [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 898f69fe-737a-44cb-9c43-965f1f8685f7] Took 1.03 seconds to deallocate network for instance. [ 636.789917] env[62569]: DEBUG nova.network.neutron [-] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.818566] env[62569]: DEBUG nova.network.neutron [req-830eaa62-9199-4c97-9d1f-5e0149d6bf06 req-a9d7c269-ec98-4e9a-812d-f4d0f4ac6deb service nova] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.200113] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.596s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.200700] env[62569]: DEBUG nova.compute.manager [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 637.203218] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.685s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.204998] env[62569]: INFO nova.compute.claims [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 637.295794] env[62569]: INFO nova.compute.manager [-] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Took 1.03 seconds to deallocate network for instance. [ 637.298132] env[62569]: DEBUG nova.compute.claims [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 637.298417] env[62569]: DEBUG oslo_concurrency.lockutils [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.321241] env[62569]: DEBUG oslo_concurrency.lockutils [req-830eaa62-9199-4c97-9d1f-5e0149d6bf06 req-a9d7c269-ec98-4e9a-812d-f4d0f4ac6deb service nova] Releasing lock "refresh_cache-21cde60b-2764-4ff7-ab5e-a17fb672351d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.321496] env[62569]: DEBUG nova.compute.manager [req-830eaa62-9199-4c97-9d1f-5e0149d6bf06 req-a9d7c269-ec98-4e9a-812d-f4d0f4ac6deb service nova] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Received event network-vif-deleted-529ffd30-aece-4729-b8da-d28c3ef597d9 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 637.711016] env[62569]: DEBUG nova.compute.utils [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 637.712661] env[62569]: DEBUG nova.compute.manager [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 637.712828] env[62569]: DEBUG nova.network.neutron [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 637.756311] env[62569]: DEBUG nova.policy [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62e586cd72a144f0a47cda57b0026bde', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '40cc8c3372fe43f7bee3ec5ccafa24db', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 637.799239] env[62569]: INFO nova.scheduler.client.report [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Deleted allocations for instance 898f69fe-737a-44cb-9c43-965f1f8685f7 [ 638.023701] env[62569]: DEBUG nova.network.neutron [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Successfully created port: d3caf564-b121-4a3b-928a-6b8bc449265b {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 638.219245] env[62569]: DEBUG nova.compute.manager [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 638.261289] env[62569]: DEBUG oslo_concurrency.lockutils [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Acquiring lock "77a1b192-6aff-4fee-93d7-57cebcdce626" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.261762] env[62569]: DEBUG oslo_concurrency.lockutils [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Lock "77a1b192-6aff-4fee-93d7-57cebcdce626" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.307705] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12f79aad-28c7-4acf-8cae-39728d024931 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "898f69fe-737a-44cb-9c43-965f1f8685f7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.473s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.624039] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d114bdf-f5f2-4f8c-87a5-5f9907661a0c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.632407] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fec46773-a76d-4794-8692-71c9429ff379 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.662983] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b48dd8f-73c3-4232-ad91-2ef2345485bb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.670547] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8517e7f5-04cc-417f-833f-00b774f4bc21 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.684804] env[62569]: DEBUG nova.compute.provider_tree [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 638.797757] env[62569]: DEBUG nova.compute.manager [req-8e0f2e1d-3b51-4d1b-b5d4-346e51f04f81 req-bdcebb9f-1bfe-472e-9463-34514306d3fa service nova] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Received event network-changed-d3caf564-b121-4a3b-928a-6b8bc449265b {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 638.797807] env[62569]: DEBUG nova.compute.manager [req-8e0f2e1d-3b51-4d1b-b5d4-346e51f04f81 req-bdcebb9f-1bfe-472e-9463-34514306d3fa service nova] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Refreshing instance network info cache due to event network-changed-d3caf564-b121-4a3b-928a-6b8bc449265b. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 638.797972] env[62569]: DEBUG oslo_concurrency.lockutils [req-8e0f2e1d-3b51-4d1b-b5d4-346e51f04f81 req-bdcebb9f-1bfe-472e-9463-34514306d3fa service nova] Acquiring lock "refresh_cache-d53a6920-d70c-40a6-a059-162a5f876b29" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.798127] env[62569]: DEBUG oslo_concurrency.lockutils [req-8e0f2e1d-3b51-4d1b-b5d4-346e51f04f81 req-bdcebb9f-1bfe-472e-9463-34514306d3fa service nova] Acquired lock "refresh_cache-d53a6920-d70c-40a6-a059-162a5f876b29" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.798288] env[62569]: DEBUG nova.network.neutron [req-8e0f2e1d-3b51-4d1b-b5d4-346e51f04f81 req-bdcebb9f-1bfe-472e-9463-34514306d3fa service nova] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Refreshing network info cache for port d3caf564-b121-4a3b-928a-6b8bc449265b {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 638.814356] env[62569]: DEBUG nova.compute.manager [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 638.970970] env[62569]: ERROR nova.compute.manager [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d3caf564-b121-4a3b-928a-6b8bc449265b, please check neutron logs for more information. [ 638.970970] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 638.970970] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 638.970970] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 638.970970] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.970970] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 638.970970] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.970970] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 638.970970] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.970970] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 638.970970] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.970970] env[62569]: ERROR nova.compute.manager raise self.value [ 638.970970] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.970970] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 638.970970] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.970970] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 638.971488] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.971488] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 638.971488] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d3caf564-b121-4a3b-928a-6b8bc449265b, please check neutron logs for more information. [ 638.971488] env[62569]: ERROR nova.compute.manager [ 638.971488] env[62569]: Traceback (most recent call last): [ 638.971488] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 638.971488] env[62569]: listener.cb(fileno) [ 638.971488] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.971488] env[62569]: result = function(*args, **kwargs) [ 638.971488] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 638.971488] env[62569]: return func(*args, **kwargs) [ 638.971488] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 638.971488] env[62569]: raise e [ 638.971488] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 638.971488] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 638.971488] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.971488] env[62569]: created_port_ids = self._update_ports_for_instance( [ 638.971488] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.971488] env[62569]: with excutils.save_and_reraise_exception(): [ 638.971488] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.971488] env[62569]: self.force_reraise() [ 638.971488] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.971488] env[62569]: raise self.value [ 638.971488] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.971488] env[62569]: updated_port = self._update_port( [ 638.971488] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.971488] env[62569]: _ensure_no_port_binding_failure(port) [ 638.971488] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.971488] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 638.972440] env[62569]: nova.exception.PortBindingFailed: Binding failed for port d3caf564-b121-4a3b-928a-6b8bc449265b, please check neutron logs for more information. [ 638.972440] env[62569]: Removing descriptor: 17 [ 639.186171] env[62569]: DEBUG nova.scheduler.client.report [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 639.231500] env[62569]: DEBUG nova.compute.manager [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 639.251711] env[62569]: DEBUG nova.virt.hardware [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 639.251955] env[62569]: DEBUG nova.virt.hardware [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 639.252126] env[62569]: DEBUG nova.virt.hardware [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 639.252314] env[62569]: DEBUG nova.virt.hardware [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 639.252462] env[62569]: DEBUG nova.virt.hardware [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 639.252607] env[62569]: DEBUG nova.virt.hardware [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 639.252986] env[62569]: DEBUG nova.virt.hardware [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 639.253241] env[62569]: DEBUG nova.virt.hardware [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 639.253355] env[62569]: DEBUG nova.virt.hardware [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 639.253521] env[62569]: DEBUG nova.virt.hardware [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 639.253695] env[62569]: DEBUG nova.virt.hardware [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 639.254564] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93e5dda3-6360-4694-a004-b6b209948370 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.262729] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bd82d91-975e-48de-a726-770185de77ee {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.276399] env[62569]: ERROR nova.compute.manager [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d3caf564-b121-4a3b-928a-6b8bc449265b, please check neutron logs for more information. [ 639.276399] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Traceback (most recent call last): [ 639.276399] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 639.276399] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] yield resources [ 639.276399] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 639.276399] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] self.driver.spawn(context, instance, image_meta, [ 639.276399] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 639.276399] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] self._vmops.spawn(context, instance, image_meta, injected_files, [ 639.276399] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 639.276399] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] vm_ref = self.build_virtual_machine(instance, [ 639.276399] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 639.276710] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] vif_infos = vmwarevif.get_vif_info(self._session, [ 639.276710] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 639.276710] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] for vif in network_info: [ 639.276710] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 639.276710] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] return self._sync_wrapper(fn, *args, **kwargs) [ 639.276710] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 639.276710] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] self.wait() [ 639.276710] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 639.276710] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] self[:] = self._gt.wait() [ 639.276710] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 639.276710] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] return self._exit_event.wait() [ 639.276710] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 639.276710] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] current.throw(*self._exc) [ 639.277092] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 639.277092] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] result = function(*args, **kwargs) [ 639.277092] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 639.277092] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] return func(*args, **kwargs) [ 639.277092] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 639.277092] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] raise e [ 639.277092] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 639.277092] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] nwinfo = self.network_api.allocate_for_instance( [ 639.277092] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 639.277092] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] created_port_ids = self._update_ports_for_instance( [ 639.277092] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 639.277092] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] with excutils.save_and_reraise_exception(): [ 639.277092] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.277411] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] self.force_reraise() [ 639.277411] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.277411] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] raise self.value [ 639.277411] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 639.277411] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] updated_port = self._update_port( [ 639.277411] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.277411] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] _ensure_no_port_binding_failure(port) [ 639.277411] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.277411] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] raise exception.PortBindingFailed(port_id=port['id']) [ 639.277411] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] nova.exception.PortBindingFailed: Binding failed for port d3caf564-b121-4a3b-928a-6b8bc449265b, please check neutron logs for more information. [ 639.277411] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] [ 639.277411] env[62569]: INFO nova.compute.manager [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Terminating instance [ 639.316916] env[62569]: DEBUG nova.network.neutron [req-8e0f2e1d-3b51-4d1b-b5d4-346e51f04f81 req-bdcebb9f-1bfe-472e-9463-34514306d3fa service nova] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.334268] env[62569]: DEBUG oslo_concurrency.lockutils [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.399186] env[62569]: DEBUG nova.network.neutron [req-8e0f2e1d-3b51-4d1b-b5d4-346e51f04f81 req-bdcebb9f-1bfe-472e-9463-34514306d3fa service nova] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.690940] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.488s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.691495] env[62569]: DEBUG nova.compute.manager [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 639.694280] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.312s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.781013] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Acquiring lock "refresh_cache-d53a6920-d70c-40a6-a059-162a5f876b29" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.901594] env[62569]: DEBUG oslo_concurrency.lockutils [req-8e0f2e1d-3b51-4d1b-b5d4-346e51f04f81 req-bdcebb9f-1bfe-472e-9463-34514306d3fa service nova] Releasing lock "refresh_cache-d53a6920-d70c-40a6-a059-162a5f876b29" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.901996] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Acquired lock "refresh_cache-d53a6920-d70c-40a6-a059-162a5f876b29" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.902216] env[62569]: DEBUG nova.network.neutron [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 640.156009] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "22094c32-5f50-4f86-a77b-cd4adcf8998a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.156247] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "22094c32-5f50-4f86-a77b-cd4adcf8998a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.199185] env[62569]: DEBUG nova.compute.utils [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 640.203829] env[62569]: DEBUG nova.compute.manager [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 640.203829] env[62569]: DEBUG nova.network.neutron [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 640.268985] env[62569]: DEBUG nova.policy [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2e2dba816cf405f9e8019701ec56c5b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2fcf4dd284d340cfb4f1f98b37aae091', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 640.430174] env[62569]: DEBUG nova.network.neutron [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 640.560647] env[62569]: DEBUG nova.network.neutron [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.629053] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19202eed-921c-4f6e-a509-4340a1c78f61 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.634986] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-281e6523-8657-4d4f-8bb1-7daaaffbebde {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.675458] env[62569]: DEBUG nova.network.neutron [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Successfully created port: c03d22b9-2146-448a-aae4-6bb8f97e497e {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 640.677753] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f1129ff-7471-463a-ac5b-77431c1c1915 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.686328] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca49a7c3-c644-4937-81fa-15e3094f443f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.699781] env[62569]: DEBUG nova.compute.provider_tree [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 640.704768] env[62569]: DEBUG nova.compute.manager [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 640.867916] env[62569]: DEBUG nova.compute.manager [req-cfafb775-8c70-473b-8818-84de05eccf9c req-44d03a5c-7217-466a-b37e-745d8503bf72 service nova] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Received event network-vif-deleted-d3caf564-b121-4a3b-928a-6b8bc449265b {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 641.065624] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Releasing lock "refresh_cache-d53a6920-d70c-40a6-a059-162a5f876b29" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.066083] env[62569]: DEBUG nova.compute.manager [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 641.066282] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 641.066573] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5ca182a5-ada9-47e6-ade6-ddeaa8708142 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.075701] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d450cb1-7e67-4b12-a872-a6645d490f63 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.100464] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d53a6920-d70c-40a6-a059-162a5f876b29 could not be found. [ 641.100710] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 641.100912] env[62569]: INFO nova.compute.manager [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Took 0.03 seconds to destroy the instance on the hypervisor. [ 641.101170] env[62569]: DEBUG oslo.service.loopingcall [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 641.101396] env[62569]: DEBUG nova.compute.manager [-] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 641.101688] env[62569]: DEBUG nova.network.neutron [-] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 641.120373] env[62569]: DEBUG nova.network.neutron [-] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.202515] env[62569]: DEBUG nova.scheduler.client.report [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 641.528589] env[62569]: ERROR nova.compute.manager [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c03d22b9-2146-448a-aae4-6bb8f97e497e, please check neutron logs for more information. [ 641.528589] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 641.528589] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 641.528589] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 641.528589] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.528589] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 641.528589] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.528589] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 641.528589] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.528589] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 641.528589] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.528589] env[62569]: ERROR nova.compute.manager raise self.value [ 641.528589] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.528589] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 641.528589] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.528589] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 641.529281] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.529281] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 641.529281] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c03d22b9-2146-448a-aae4-6bb8f97e497e, please check neutron logs for more information. [ 641.529281] env[62569]: ERROR nova.compute.manager [ 641.529281] env[62569]: Traceback (most recent call last): [ 641.529281] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 641.529281] env[62569]: listener.cb(fileno) [ 641.529281] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.529281] env[62569]: result = function(*args, **kwargs) [ 641.529281] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 641.529281] env[62569]: return func(*args, **kwargs) [ 641.529281] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 641.529281] env[62569]: raise e [ 641.529281] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 641.529281] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 641.529281] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.529281] env[62569]: created_port_ids = self._update_ports_for_instance( [ 641.529281] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.529281] env[62569]: with excutils.save_and_reraise_exception(): [ 641.529281] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.529281] env[62569]: self.force_reraise() [ 641.529281] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.529281] env[62569]: raise self.value [ 641.529281] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.529281] env[62569]: updated_port = self._update_port( [ 641.529281] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.529281] env[62569]: _ensure_no_port_binding_failure(port) [ 641.529281] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.529281] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 641.530024] env[62569]: nova.exception.PortBindingFailed: Binding failed for port c03d22b9-2146-448a-aae4-6bb8f97e497e, please check neutron logs for more information. [ 641.530024] env[62569]: Removing descriptor: 17 [ 641.622650] env[62569]: DEBUG nova.network.neutron [-] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.709035] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.015s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.709720] env[62569]: ERROR nova.compute.manager [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a5315c23-9de0-4015-ae25-df72c2a8557e, please check neutron logs for more information. [ 641.709720] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Traceback (most recent call last): [ 641.709720] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 641.709720] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] self.driver.spawn(context, instance, image_meta, [ 641.709720] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 641.709720] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 641.709720] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 641.709720] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] vm_ref = self.build_virtual_machine(instance, [ 641.709720] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 641.709720] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] vif_infos = vmwarevif.get_vif_info(self._session, [ 641.709720] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 641.710036] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] for vif in network_info: [ 641.710036] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 641.710036] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] return self._sync_wrapper(fn, *args, **kwargs) [ 641.710036] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 641.710036] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] self.wait() [ 641.710036] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 641.710036] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] self[:] = self._gt.wait() [ 641.710036] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 641.710036] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] return self._exit_event.wait() [ 641.710036] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 641.710036] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] current.throw(*self._exc) [ 641.710036] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.710036] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] result = function(*args, **kwargs) [ 641.710337] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 641.710337] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] return func(*args, **kwargs) [ 641.710337] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 641.710337] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] raise e [ 641.710337] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 641.710337] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] nwinfo = self.network_api.allocate_for_instance( [ 641.710337] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.710337] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] created_port_ids = self._update_ports_for_instance( [ 641.710337] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.710337] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] with excutils.save_and_reraise_exception(): [ 641.710337] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.710337] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] self.force_reraise() [ 641.710337] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.710652] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] raise self.value [ 641.710652] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.710652] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] updated_port = self._update_port( [ 641.710652] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.710652] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] _ensure_no_port_binding_failure(port) [ 641.710652] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.710652] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] raise exception.PortBindingFailed(port_id=port['id']) [ 641.710652] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] nova.exception.PortBindingFailed: Binding failed for port a5315c23-9de0-4015-ae25-df72c2a8557e, please check neutron logs for more information. [ 641.710652] env[62569]: ERROR nova.compute.manager [instance: 94818059-cf10-4dca-a711-fd5809583ce9] [ 641.710652] env[62569]: DEBUG nova.compute.utils [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Binding failed for port a5315c23-9de0-4015-ae25-df72c2a8557e, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 641.712125] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.837s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.715384] env[62569]: DEBUG nova.compute.manager [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 641.718031] env[62569]: DEBUG nova.compute.manager [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Build of instance 94818059-cf10-4dca-a711-fd5809583ce9 was re-scheduled: Binding failed for port a5315c23-9de0-4015-ae25-df72c2a8557e, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 641.718448] env[62569]: DEBUG nova.compute.manager [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 641.718543] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Acquiring lock "refresh_cache-94818059-cf10-4dca-a711-fd5809583ce9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.719340] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Acquired lock "refresh_cache-94818059-cf10-4dca-a711-fd5809583ce9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.719340] env[62569]: DEBUG nova.network.neutron [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 641.741791] env[62569]: DEBUG nova.virt.hardware [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 641.742048] env[62569]: DEBUG nova.virt.hardware [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 641.742213] env[62569]: DEBUG nova.virt.hardware [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 641.742396] env[62569]: DEBUG nova.virt.hardware [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 641.742541] env[62569]: DEBUG nova.virt.hardware [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 641.742688] env[62569]: DEBUG nova.virt.hardware [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 641.742889] env[62569]: DEBUG nova.virt.hardware [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 641.743165] env[62569]: DEBUG nova.virt.hardware [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 641.743230] env[62569]: DEBUG nova.virt.hardware [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 641.743383] env[62569]: DEBUG nova.virt.hardware [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 641.744116] env[62569]: DEBUG nova.virt.hardware [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 641.744644] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81022d59-51f0-47ea-8e04-6de4398643e6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.753974] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62432611-d108-43ae-88c3-f50dc2181866 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.770389] env[62569]: ERROR nova.compute.manager [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c03d22b9-2146-448a-aae4-6bb8f97e497e, please check neutron logs for more information. [ 641.770389] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] Traceback (most recent call last): [ 641.770389] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 641.770389] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] yield resources [ 641.770389] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 641.770389] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] self.driver.spawn(context, instance, image_meta, [ 641.770389] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 641.770389] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] self._vmops.spawn(context, instance, image_meta, injected_files, [ 641.770389] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 641.770389] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] vm_ref = self.build_virtual_machine(instance, [ 641.770389] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 641.770752] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] vif_infos = vmwarevif.get_vif_info(self._session, [ 641.770752] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 641.770752] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] for vif in network_info: [ 641.770752] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 641.770752] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] return self._sync_wrapper(fn, *args, **kwargs) [ 641.770752] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 641.770752] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] self.wait() [ 641.770752] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 641.770752] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] self[:] = self._gt.wait() [ 641.770752] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 641.770752] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] return self._exit_event.wait() [ 641.770752] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 641.770752] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] current.throw(*self._exc) [ 641.771107] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.771107] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] result = function(*args, **kwargs) [ 641.771107] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 641.771107] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] return func(*args, **kwargs) [ 641.771107] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 641.771107] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] raise e [ 641.771107] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 641.771107] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] nwinfo = self.network_api.allocate_for_instance( [ 641.771107] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.771107] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] created_port_ids = self._update_ports_for_instance( [ 641.771107] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.771107] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] with excutils.save_and_reraise_exception(): [ 641.771107] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.771577] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] self.force_reraise() [ 641.771577] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.771577] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] raise self.value [ 641.771577] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.771577] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] updated_port = self._update_port( [ 641.771577] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.771577] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] _ensure_no_port_binding_failure(port) [ 641.771577] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.771577] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] raise exception.PortBindingFailed(port_id=port['id']) [ 641.771577] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] nova.exception.PortBindingFailed: Binding failed for port c03d22b9-2146-448a-aae4-6bb8f97e497e, please check neutron logs for more information. [ 641.771577] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] [ 641.771577] env[62569]: INFO nova.compute.manager [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Terminating instance [ 642.125662] env[62569]: INFO nova.compute.manager [-] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Took 1.02 seconds to deallocate network for instance. [ 642.128068] env[62569]: DEBUG nova.compute.claims [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 642.128247] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.242520] env[62569]: DEBUG nova.network.neutron [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.276293] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Acquiring lock "refresh_cache-e53104ad-35c7-49cb-b8de-035006148c52" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.276470] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Acquired lock "refresh_cache-e53104ad-35c7-49cb-b8de-035006148c52" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.276638] env[62569]: DEBUG nova.network.neutron [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 642.330570] env[62569]: DEBUG nova.network.neutron [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.621841] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f76096a-a9eb-423b-bc0d-041c302ac74b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.629229] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-055072e9-a581-444c-bc24-2729a145cf61 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.658430] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ae4ba6f-75fb-4a4e-a41e-a5ed4538ecbd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.665717] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-543d8169-66ae-407c-92d8-2cd15ea9fc43 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.679279] env[62569]: DEBUG nova.compute.provider_tree [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 642.793979] env[62569]: DEBUG nova.network.neutron [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.833311] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Releasing lock "refresh_cache-94818059-cf10-4dca-a711-fd5809583ce9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.833541] env[62569]: DEBUG nova.compute.manager [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 642.833722] env[62569]: DEBUG nova.compute.manager [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 642.833890] env[62569]: DEBUG nova.network.neutron [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 642.867176] env[62569]: DEBUG nova.network.neutron [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.873243] env[62569]: DEBUG nova.network.neutron [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.895438] env[62569]: DEBUG nova.compute.manager [req-0a3b4cec-d7a6-41ff-94f2-a19d740bd9f3 req-4459fc73-5461-4eff-9a15-430a03549b20 service nova] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Received event network-changed-c03d22b9-2146-448a-aae4-6bb8f97e497e {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 642.895701] env[62569]: DEBUG nova.compute.manager [req-0a3b4cec-d7a6-41ff-94f2-a19d740bd9f3 req-4459fc73-5461-4eff-9a15-430a03549b20 service nova] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Refreshing instance network info cache due to event network-changed-c03d22b9-2146-448a-aae4-6bb8f97e497e. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 642.895845] env[62569]: DEBUG oslo_concurrency.lockutils [req-0a3b4cec-d7a6-41ff-94f2-a19d740bd9f3 req-4459fc73-5461-4eff-9a15-430a03549b20 service nova] Acquiring lock "refresh_cache-e53104ad-35c7-49cb-b8de-035006148c52" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.182341] env[62569]: DEBUG nova.scheduler.client.report [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 643.369937] env[62569]: DEBUG nova.network.neutron [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.375842] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Releasing lock "refresh_cache-e53104ad-35c7-49cb-b8de-035006148c52" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.376252] env[62569]: DEBUG nova.compute.manager [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 643.376442] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 643.376742] env[62569]: DEBUG oslo_concurrency.lockutils [req-0a3b4cec-d7a6-41ff-94f2-a19d740bd9f3 req-4459fc73-5461-4eff-9a15-430a03549b20 service nova] Acquired lock "refresh_cache-e53104ad-35c7-49cb-b8de-035006148c52" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.376913] env[62569]: DEBUG nova.network.neutron [req-0a3b4cec-d7a6-41ff-94f2-a19d740bd9f3 req-4459fc73-5461-4eff-9a15-430a03549b20 service nova] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Refreshing network info cache for port c03d22b9-2146-448a-aae4-6bb8f97e497e {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 643.378225] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-63e1a854-400f-4607-a805-ffc27bffc890 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.388304] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4801224e-e3d3-458a-b550-132cf13d1c39 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.409968] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e53104ad-35c7-49cb-b8de-035006148c52 could not be found. [ 643.410213] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 643.410391] env[62569]: INFO nova.compute.manager [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Took 0.03 seconds to destroy the instance on the hypervisor. [ 643.410675] env[62569]: DEBUG oslo.service.loopingcall [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 643.410950] env[62569]: DEBUG nova.compute.manager [-] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 643.411058] env[62569]: DEBUG nova.network.neutron [-] [instance: e53104ad-35c7-49cb-b8de-035006148c52] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 643.426512] env[62569]: DEBUG nova.network.neutron [-] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.687196] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.975s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.687828] env[62569]: ERROR nova.compute.manager [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 78e8ea53-39d8-4665-a85e-497722b2d3c3, please check neutron logs for more information. [ 643.687828] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Traceback (most recent call last): [ 643.687828] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 643.687828] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] self.driver.spawn(context, instance, image_meta, [ 643.687828] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 643.687828] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] self._vmops.spawn(context, instance, image_meta, injected_files, [ 643.687828] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 643.687828] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] vm_ref = self.build_virtual_machine(instance, [ 643.687828] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 643.687828] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] vif_infos = vmwarevif.get_vif_info(self._session, [ 643.687828] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 643.688280] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] for vif in network_info: [ 643.688280] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 643.688280] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] return self._sync_wrapper(fn, *args, **kwargs) [ 643.688280] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 643.688280] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] self.wait() [ 643.688280] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 643.688280] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] self[:] = self._gt.wait() [ 643.688280] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 643.688280] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] return self._exit_event.wait() [ 643.688280] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 643.688280] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] result = hub.switch() [ 643.688280] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 643.688280] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] return self.greenlet.switch() [ 643.688798] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 643.688798] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] result = function(*args, **kwargs) [ 643.688798] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 643.688798] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] return func(*args, **kwargs) [ 643.688798] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 643.688798] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] raise e [ 643.688798] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 643.688798] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] nwinfo = self.network_api.allocate_for_instance( [ 643.688798] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 643.688798] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] created_port_ids = self._update_ports_for_instance( [ 643.688798] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 643.688798] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] with excutils.save_and_reraise_exception(): [ 643.688798] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 643.689361] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] self.force_reraise() [ 643.689361] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 643.689361] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] raise self.value [ 643.689361] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 643.689361] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] updated_port = self._update_port( [ 643.689361] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 643.689361] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] _ensure_no_port_binding_failure(port) [ 643.689361] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 643.689361] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] raise exception.PortBindingFailed(port_id=port['id']) [ 643.689361] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] nova.exception.PortBindingFailed: Binding failed for port 78e8ea53-39d8-4665-a85e-497722b2d3c3, please check neutron logs for more information. [ 643.689361] env[62569]: ERROR nova.compute.manager [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] [ 643.689724] env[62569]: DEBUG nova.compute.utils [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Binding failed for port 78e8ea53-39d8-4665-a85e-497722b2d3c3, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 643.689763] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 16.694s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.689950] env[62569]: DEBUG nova.objects.instance [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62569) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 643.692528] env[62569]: DEBUG nova.compute.manager [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Build of instance c29ad19e-91a5-4c3d-912b-bd7be4c42016 was re-scheduled: Binding failed for port 78e8ea53-39d8-4665-a85e-497722b2d3c3, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 643.692959] env[62569]: DEBUG nova.compute.manager [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 643.693199] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Acquiring lock "refresh_cache-c29ad19e-91a5-4c3d-912b-bd7be4c42016" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.693353] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Acquired lock "refresh_cache-c29ad19e-91a5-4c3d-912b-bd7be4c42016" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.693506] env[62569]: DEBUG nova.network.neutron [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 643.873699] env[62569]: INFO nova.compute.manager [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] [instance: 94818059-cf10-4dca-a711-fd5809583ce9] Took 1.04 seconds to deallocate network for instance. [ 643.896492] env[62569]: DEBUG nova.network.neutron [req-0a3b4cec-d7a6-41ff-94f2-a19d740bd9f3 req-4459fc73-5461-4eff-9a15-430a03549b20 service nova] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.928899] env[62569]: DEBUG nova.network.neutron [-] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.967467] env[62569]: DEBUG nova.network.neutron [req-0a3b4cec-d7a6-41ff-94f2-a19d740bd9f3 req-4459fc73-5461-4eff-9a15-430a03549b20 service nova] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.213099] env[62569]: DEBUG nova.network.neutron [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 644.280279] env[62569]: DEBUG nova.network.neutron [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.430935] env[62569]: INFO nova.compute.manager [-] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Took 1.02 seconds to deallocate network for instance. [ 644.433364] env[62569]: DEBUG nova.compute.claims [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 644.433541] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.469989] env[62569]: DEBUG oslo_concurrency.lockutils [req-0a3b4cec-d7a6-41ff-94f2-a19d740bd9f3 req-4459fc73-5461-4eff-9a15-430a03549b20 service nova] Releasing lock "refresh_cache-e53104ad-35c7-49cb-b8de-035006148c52" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.470281] env[62569]: DEBUG nova.compute.manager [req-0a3b4cec-d7a6-41ff-94f2-a19d740bd9f3 req-4459fc73-5461-4eff-9a15-430a03549b20 service nova] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Received event network-vif-deleted-c03d22b9-2146-448a-aae4-6bb8f97e497e {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 644.703037] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a5c4a181-3505-42c5-bb65-f0a629b598ff tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.704169] env[62569]: DEBUG oslo_concurrency.lockutils [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.166s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.705612] env[62569]: INFO nova.compute.claims [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 644.782995] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Releasing lock "refresh_cache-c29ad19e-91a5-4c3d-912b-bd7be4c42016" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.783262] env[62569]: DEBUG nova.compute.manager [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 644.783442] env[62569]: DEBUG nova.compute.manager [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 644.783606] env[62569]: DEBUG nova.network.neutron [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 644.798350] env[62569]: DEBUG nova.network.neutron [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 644.907933] env[62569]: INFO nova.scheduler.client.report [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Deleted allocations for instance 94818059-cf10-4dca-a711-fd5809583ce9 [ 645.300584] env[62569]: DEBUG nova.network.neutron [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.416742] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f27e7baa-574b-436b-a5bc-0e2094f45682 tempest-ServersTestBootFromVolume-208288499 tempest-ServersTestBootFromVolume-208288499-project-member] Lock "94818059-cf10-4dca-a711-fd5809583ce9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.151s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.803495] env[62569]: INFO nova.compute.manager [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] [instance: c29ad19e-91a5-4c3d-912b-bd7be4c42016] Took 1.02 seconds to deallocate network for instance. [ 645.919740] env[62569]: DEBUG nova.compute.manager [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 646.087051] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93870870-26a3-4d39-83fc-cf984c1e1c7e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.093309] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2b37e6b-28bb-434a-8ca5-5f5ee4aeeaad {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.122370] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f69bf10-ff18-4144-a734-28aa31bc9073 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.129650] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-849bb021-0d38-4fc8-bd2d-b2bd5ba202b6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.143734] env[62569]: DEBUG nova.compute.provider_tree [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 646.443695] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.650637] env[62569]: DEBUG nova.scheduler.client.report [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 646.838546] env[62569]: INFO nova.scheduler.client.report [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Deleted allocations for instance c29ad19e-91a5-4c3d-912b-bd7be4c42016 [ 647.158297] env[62569]: DEBUG oslo_concurrency.lockutils [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.454s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.158819] env[62569]: DEBUG nova.compute.manager [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 647.161571] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 17.044s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.351143] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e55fcc06-be05-4c5f-96c3-55cfb4eb8f69 tempest-ServerAddressesNegativeTestJSON-2055780355 tempest-ServerAddressesNegativeTestJSON-2055780355-project-member] Lock "c29ad19e-91a5-4c3d-912b-bd7be4c42016" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.652s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.665806] env[62569]: DEBUG nova.compute.utils [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 647.673017] env[62569]: DEBUG nova.compute.manager [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 647.673214] env[62569]: DEBUG nova.network.neutron [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 647.742084] env[62569]: DEBUG nova.policy [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '840341f9ca2f4eb5b73a926798c5be8f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd38467713b6340ab9ca2e0d8cbd799a6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 647.854770] env[62569]: DEBUG nova.compute.manager [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 648.156445] env[62569]: DEBUG nova.network.neutron [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Successfully created port: 30a8edfc-9961-47b8-910a-626126d325e7 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 648.173648] env[62569]: DEBUG nova.compute.manager [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 648.212874] env[62569]: WARNING nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 4b9e4f73-4682-4170-b71e-1d3d1d860577 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 648.213058] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 418c648a-06be-4e97-b138-45c65addf5cf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 648.213191] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 92666c7b-7a1d-4d49-931f-e84cd2486871 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 648.213308] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 21cde60b-2764-4ff7-ab5e-a17fb672351d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 648.213420] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance d53a6920-d70c-40a6-a059-162a5f876b29 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 648.213532] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance e53104ad-35c7-49cb-b8de-035006148c52 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 648.213641] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 7cb248f2-2db2-4f62-b2d1-332488170c9b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 648.378787] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.716573] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 695022e4-bb58-44d7-87e5-58bbd577ec7b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.117256] env[62569]: DEBUG nova.compute.manager [req-20b5cfc9-32b1-4c9d-9ce8-f19473b53000 req-ed2a405b-2aa5-434f-81a7-b9588cbe4014 service nova] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Received event network-changed-30a8edfc-9961-47b8-910a-626126d325e7 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 649.117518] env[62569]: DEBUG nova.compute.manager [req-20b5cfc9-32b1-4c9d-9ce8-f19473b53000 req-ed2a405b-2aa5-434f-81a7-b9588cbe4014 service nova] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Refreshing instance network info cache due to event network-changed-30a8edfc-9961-47b8-910a-626126d325e7. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 649.117655] env[62569]: DEBUG oslo_concurrency.lockutils [req-20b5cfc9-32b1-4c9d-9ce8-f19473b53000 req-ed2a405b-2aa5-434f-81a7-b9588cbe4014 service nova] Acquiring lock "refresh_cache-7cb248f2-2db2-4f62-b2d1-332488170c9b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 649.117803] env[62569]: DEBUG oslo_concurrency.lockutils [req-20b5cfc9-32b1-4c9d-9ce8-f19473b53000 req-ed2a405b-2aa5-434f-81a7-b9588cbe4014 service nova] Acquired lock "refresh_cache-7cb248f2-2db2-4f62-b2d1-332488170c9b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.117953] env[62569]: DEBUG nova.network.neutron [req-20b5cfc9-32b1-4c9d-9ce8-f19473b53000 req-ed2a405b-2aa5-434f-81a7-b9588cbe4014 service nova] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Refreshing network info cache for port 30a8edfc-9961-47b8-910a-626126d325e7 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 649.186280] env[62569]: DEBUG nova.compute.manager [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 649.217668] env[62569]: DEBUG nova.virt.hardware [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:32:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='e5d62faf-1856-4402-8065-2ffd5b34dbf5',id=35,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-870974799',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 649.217913] env[62569]: DEBUG nova.virt.hardware [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 649.218086] env[62569]: DEBUG nova.virt.hardware [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 649.218274] env[62569]: DEBUG nova.virt.hardware [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 649.218421] env[62569]: DEBUG nova.virt.hardware [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 649.218567] env[62569]: DEBUG nova.virt.hardware [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 649.218778] env[62569]: DEBUG nova.virt.hardware [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 649.218938] env[62569]: DEBUG nova.virt.hardware [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 649.219190] env[62569]: DEBUG nova.virt.hardware [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 649.219366] env[62569]: DEBUG nova.virt.hardware [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 649.219538] env[62569]: DEBUG nova.virt.hardware [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 649.220264] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.223549] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8aad13f-d528-409c-ad74-8b222132a459 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.233690] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c33d139c-7b29-4935-b405-dee005620f9a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.296857] env[62569]: ERROR nova.compute.manager [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 30a8edfc-9961-47b8-910a-626126d325e7, please check neutron logs for more information. [ 649.296857] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 649.296857] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 649.296857] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 649.296857] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.296857] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 649.296857] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.296857] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 649.296857] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.296857] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 649.296857] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.296857] env[62569]: ERROR nova.compute.manager raise self.value [ 649.296857] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.296857] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 649.296857] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.296857] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 649.297467] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.297467] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 649.297467] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 30a8edfc-9961-47b8-910a-626126d325e7, please check neutron logs for more information. [ 649.297467] env[62569]: ERROR nova.compute.manager [ 649.297467] env[62569]: Traceback (most recent call last): [ 649.297467] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 649.297467] env[62569]: listener.cb(fileno) [ 649.297467] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 649.297467] env[62569]: result = function(*args, **kwargs) [ 649.297467] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 649.297467] env[62569]: return func(*args, **kwargs) [ 649.297467] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 649.297467] env[62569]: raise e [ 649.297467] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 649.297467] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 649.297467] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.297467] env[62569]: created_port_ids = self._update_ports_for_instance( [ 649.297467] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.297467] env[62569]: with excutils.save_and_reraise_exception(): [ 649.297467] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.297467] env[62569]: self.force_reraise() [ 649.297467] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.297467] env[62569]: raise self.value [ 649.297467] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.297467] env[62569]: updated_port = self._update_port( [ 649.297467] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.297467] env[62569]: _ensure_no_port_binding_failure(port) [ 649.297467] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.297467] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 649.298287] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 30a8edfc-9961-47b8-910a-626126d325e7, please check neutron logs for more information. [ 649.298287] env[62569]: Removing descriptor: 17 [ 649.298287] env[62569]: ERROR nova.compute.manager [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 30a8edfc-9961-47b8-910a-626126d325e7, please check neutron logs for more information. [ 649.298287] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Traceback (most recent call last): [ 649.298287] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 649.298287] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] yield resources [ 649.298287] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 649.298287] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] self.driver.spawn(context, instance, image_meta, [ 649.298287] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 649.298287] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 649.298287] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 649.298287] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] vm_ref = self.build_virtual_machine(instance, [ 649.298637] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 649.298637] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] vif_infos = vmwarevif.get_vif_info(self._session, [ 649.298637] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 649.298637] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] for vif in network_info: [ 649.298637] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 649.298637] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] return self._sync_wrapper(fn, *args, **kwargs) [ 649.298637] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 649.298637] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] self.wait() [ 649.298637] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 649.298637] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] self[:] = self._gt.wait() [ 649.298637] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 649.298637] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] return self._exit_event.wait() [ 649.298637] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 649.299059] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] result = hub.switch() [ 649.299059] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 649.299059] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] return self.greenlet.switch() [ 649.299059] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 649.299059] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] result = function(*args, **kwargs) [ 649.299059] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 649.299059] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] return func(*args, **kwargs) [ 649.299059] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 649.299059] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] raise e [ 649.299059] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 649.299059] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] nwinfo = self.network_api.allocate_for_instance( [ 649.299059] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.299059] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] created_port_ids = self._update_ports_for_instance( [ 649.299391] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.299391] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] with excutils.save_and_reraise_exception(): [ 649.299391] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.299391] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] self.force_reraise() [ 649.299391] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.299391] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] raise self.value [ 649.299391] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.299391] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] updated_port = self._update_port( [ 649.299391] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.299391] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] _ensure_no_port_binding_failure(port) [ 649.299391] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.299391] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] raise exception.PortBindingFailed(port_id=port['id']) [ 649.299761] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] nova.exception.PortBindingFailed: Binding failed for port 30a8edfc-9961-47b8-910a-626126d325e7, please check neutron logs for more information. [ 649.299761] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] [ 649.299761] env[62569]: INFO nova.compute.manager [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Terminating instance [ 649.638464] env[62569]: DEBUG nova.network.neutron [req-20b5cfc9-32b1-4c9d-9ce8-f19473b53000 req-ed2a405b-2aa5-434f-81a7-b9588cbe4014 service nova] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.723605] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.725643] env[62569]: DEBUG nova.network.neutron [req-20b5cfc9-32b1-4c9d-9ce8-f19473b53000 req-ed2a405b-2aa5-434f-81a7-b9588cbe4014 service nova] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.803944] env[62569]: DEBUG oslo_concurrency.lockutils [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Acquiring lock "refresh_cache-7cb248f2-2db2-4f62-b2d1-332488170c9b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.228793] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance c3f25b97-0e1e-4e78-94f0-e8087d1b55c1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 650.230294] env[62569]: DEBUG oslo_concurrency.lockutils [req-20b5cfc9-32b1-4c9d-9ce8-f19473b53000 req-ed2a405b-2aa5-434f-81a7-b9588cbe4014 service nova] Releasing lock "refresh_cache-7cb248f2-2db2-4f62-b2d1-332488170c9b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.230834] env[62569]: DEBUG oslo_concurrency.lockutils [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Acquired lock "refresh_cache-7cb248f2-2db2-4f62-b2d1-332488170c9b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.231019] env[62569]: DEBUG nova.network.neutron [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 650.731817] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance d3274d59-6cf7-4e86-90f9-ffea49ad5342 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 650.761619] env[62569]: DEBUG nova.network.neutron [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.848405] env[62569]: DEBUG nova.network.neutron [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.216718] env[62569]: DEBUG nova.compute.manager [req-016fadd5-a5b4-4c0c-b9dd-e36f291e9c31 req-029efd3c-e262-4b1e-99aa-368227c33d4a service nova] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Received event network-vif-deleted-30a8edfc-9961-47b8-910a-626126d325e7 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 651.239509] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 4f45ccc2-0352-4c25-9a92-aee8f051c0b5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 651.349680] env[62569]: DEBUG oslo_concurrency.lockutils [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Releasing lock "refresh_cache-7cb248f2-2db2-4f62-b2d1-332488170c9b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.350181] env[62569]: DEBUG nova.compute.manager [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 651.350379] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 651.350685] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-769e586e-b76d-410f-9581-2533804033b8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.363305] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d29546a0-6b30-480c-a5f3-60f3879fb006 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.385155] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7cb248f2-2db2-4f62-b2d1-332488170c9b could not be found. [ 651.385356] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 651.385487] env[62569]: INFO nova.compute.manager [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 651.385778] env[62569]: DEBUG oslo.service.loopingcall [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 651.386081] env[62569]: DEBUG nova.compute.manager [-] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 651.386081] env[62569]: DEBUG nova.network.neutron [-] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 651.407695] env[62569]: DEBUG nova.network.neutron [-] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.744300] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance bda4e4c4-b1e2-4441-8aa5-6d1b5975647c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 651.912474] env[62569]: DEBUG nova.network.neutron [-] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.251628] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance a930be54-d8c9-4407-9d49-f067defc65e9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 652.416727] env[62569]: INFO nova.compute.manager [-] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Took 1.03 seconds to deallocate network for instance. [ 652.426357] env[62569]: DEBUG nova.compute.claims [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 652.426357] env[62569]: DEBUG oslo_concurrency.lockutils [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.756855] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 34173a5f-8e4d-4646-b3e3-d537ffe752bb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 652.781936] env[62569]: DEBUG oslo_concurrency.lockutils [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Acquiring lock "4feb7b78-9f7c-4e64-b0a7-870ed73adf97" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.782502] env[62569]: DEBUG oslo_concurrency.lockutils [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Lock "4feb7b78-9f7c-4e64-b0a7-870ed73adf97" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.266932] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 68218523-ae44-4332-8b20-b23559304d60 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 653.772682] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 311e46eb-85ca-4262-a0a4-21eb073e0dc4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 654.279178] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 17f0965b-edf8-4590-9402-2654d5d19f4e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 654.781156] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 655.284167] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 655.787664] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 93c8445d-689c-4f77-836f-95eafafbb6a2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 656.292422] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 3f504d5c-3431-4a80-a7d9-e52af01cf24e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 656.746678] env[62569]: DEBUG oslo_concurrency.lockutils [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Acquiring lock "0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.746921] env[62569]: DEBUG oslo_concurrency.lockutils [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Lock "0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.795531] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance a0b406c3-9466-41bd-9de1-e675cab2ceef has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 657.299841] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 94392d48-223d-4205-9627-25488a468769 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 657.803589] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 015140aa-8336-40e2-856c-723277a48f6e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 658.306044] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance c515d85e-bcb5-4bac-bacb-1e558f38171f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 658.809402] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance fd763e12-5b29-48f8-8256-cb6205e0d119 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 659.314870] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 77a1b192-6aff-4fee-93d7-57cebcdce626 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 659.817845] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 22094c32-5f50-4f86-a77b-cd4adcf8998a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 659.818214] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=62569) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 659.818408] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=62569) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 660.168171] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c94f9c8d-aac8-45a9-9e4d-abfd6a7a1d43 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.176760] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f01ef6f1-6485-4d32-bab3-adec574745eb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.205307] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4e3312f-5041-4c55-af14-839a19bd4719 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.212790] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8f5d462-2f7a-452e-9a1c-3ebc07a03345 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.226337] env[62569]: DEBUG nova.compute.provider_tree [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.729992] env[62569]: DEBUG nova.scheduler.client.report [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 661.234946] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62569) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 661.235222] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 14.074s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.235511] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.097s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.075568] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1a4e76d-bcdb-47ec-ab0c-98b81a512b97 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.082998] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5643cd6-b17a-4545-8024-54845b627e8a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.111648] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b3f6e1-86ed-4698-a859-b6d0cf157dcc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.118224] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc4c97ea-c7f2-4cbe-8bed-4d42bbb402c7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.130363] env[62569]: DEBUG nova.compute.provider_tree [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.633285] env[62569]: DEBUG nova.scheduler.client.report [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 663.138258] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.903s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.138912] env[62569]: ERROR nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 12465efc-17a2-432b-a4c4-ff14cf34a080, please check neutron logs for more information. [ 663.138912] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Traceback (most recent call last): [ 663.138912] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 663.138912] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] self.driver.spawn(context, instance, image_meta, [ 663.138912] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 663.138912] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 663.138912] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 663.138912] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] vm_ref = self.build_virtual_machine(instance, [ 663.138912] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 663.138912] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] vif_infos = vmwarevif.get_vif_info(self._session, [ 663.138912] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 663.139249] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] for vif in network_info: [ 663.139249] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 663.139249] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] return self._sync_wrapper(fn, *args, **kwargs) [ 663.139249] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 663.139249] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] self.wait() [ 663.139249] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 663.139249] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] self[:] = self._gt.wait() [ 663.139249] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 663.139249] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] return self._exit_event.wait() [ 663.139249] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 663.139249] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] current.throw(*self._exc) [ 663.139249] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.139249] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] result = function(*args, **kwargs) [ 663.139559] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 663.139559] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] return func(*args, **kwargs) [ 663.139559] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 663.139559] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] raise e [ 663.139559] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 663.139559] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] nwinfo = self.network_api.allocate_for_instance( [ 663.139559] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 663.139559] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] created_port_ids = self._update_ports_for_instance( [ 663.139559] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 663.139559] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] with excutils.save_and_reraise_exception(): [ 663.139559] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.139559] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] self.force_reraise() [ 663.139559] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.139988] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] raise self.value [ 663.139988] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 663.139988] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] updated_port = self._update_port( [ 663.139988] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.139988] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] _ensure_no_port_binding_failure(port) [ 663.139988] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.139988] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] raise exception.PortBindingFailed(port_id=port['id']) [ 663.139988] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] nova.exception.PortBindingFailed: Binding failed for port 12465efc-17a2-432b-a4c4-ff14cf34a080, please check neutron logs for more information. [ 663.139988] env[62569]: ERROR nova.compute.manager [instance: 418c648a-06be-4e97-b138-45c65addf5cf] [ 663.139988] env[62569]: DEBUG nova.compute.utils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Binding failed for port 12465efc-17a2-432b-a4c4-ff14cf34a080, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 663.141084] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.263s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.141084] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.142964] env[62569]: DEBUG oslo_concurrency.lockutils [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.963s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.144810] env[62569]: INFO nova.compute.claims [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 663.147515] env[62569]: DEBUG nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Build of instance 418c648a-06be-4e97-b138-45c65addf5cf was re-scheduled: Binding failed for port 12465efc-17a2-432b-a4c4-ff14cf34a080, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 663.147993] env[62569]: DEBUG nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 663.148201] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "refresh_cache-418c648a-06be-4e97-b138-45c65addf5cf" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.148346] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquired lock "refresh_cache-418c648a-06be-4e97-b138-45c65addf5cf" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.148502] env[62569]: DEBUG nova.network.neutron [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 663.167308] env[62569]: INFO nova.scheduler.client.report [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Deleted allocations for instance 4b9e4f73-4682-4170-b71e-1d3d1d860577 [ 663.674547] env[62569]: DEBUG nova.network.neutron [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.678139] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cadc8a9d-4478-4a0e-b162-6d97ce10d6df tempest-ServerShowV257Test-1457122217 tempest-ServerShowV257Test-1457122217-project-member] Lock "4b9e4f73-4682-4170-b71e-1d3d1d860577" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.981s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.803360] env[62569]: DEBUG nova.network.neutron [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.306810] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Releasing lock "refresh_cache-418c648a-06be-4e97-b138-45c65addf5cf" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.307727] env[62569]: DEBUG nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 664.307727] env[62569]: DEBUG nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 664.307727] env[62569]: DEBUG nova.network.neutron [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 664.325153] env[62569]: DEBUG nova.network.neutron [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.530026] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90c00527-f56a-4103-a824-eddcd68f6f14 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.537734] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da159a0c-ef23-4b2f-8477-3f406c114794 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.571576] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a340c599-606a-4e82-b199-63b46bdcfc46 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.580289] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc8ba9b-a852-4367-946b-2cf3f7841c00 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.596107] env[62569]: DEBUG nova.compute.provider_tree [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 664.828256] env[62569]: DEBUG nova.network.neutron [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.099168] env[62569]: DEBUG nova.scheduler.client.report [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 665.333952] env[62569]: INFO nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 418c648a-06be-4e97-b138-45c65addf5cf] Took 1.02 seconds to deallocate network for instance. [ 665.610974] env[62569]: DEBUG oslo_concurrency.lockutils [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.468s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.611553] env[62569]: DEBUG nova.compute.manager [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 665.615406] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.737s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.123569] env[62569]: DEBUG nova.compute.utils [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 666.128975] env[62569]: DEBUG nova.compute.manager [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 666.129305] env[62569]: DEBUG nova.network.neutron [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 666.179450] env[62569]: DEBUG nova.policy [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f6f94ef6125e4f799800d37471b9eb6f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8c69ac971dbd419f870a9fa68850cb1c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 666.368031] env[62569]: INFO nova.scheduler.client.report [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Deleted allocations for instance 418c648a-06be-4e97-b138-45c65addf5cf [ 666.498560] env[62569]: DEBUG nova.network.neutron [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Successfully created port: a0f99f88-8014-4300-863a-38e81bc640d2 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 666.575407] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-325cdc3a-7f2b-4b70-bdac-0dd63d27b8ae {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.583594] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a41f3ee-ce90-4099-973d-d404e2af6710 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.621117] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50006ca5-637a-437c-9269-f2d69bc169bc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.626130] env[62569]: DEBUG nova.compute.manager [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 666.633043] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc60427a-0b56-4abb-a3fd-8db801dd788d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.651899] env[62569]: DEBUG nova.compute.provider_tree [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 666.886885] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "418c648a-06be-4e97-b138-45c65addf5cf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.769s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.155456] env[62569]: DEBUG nova.scheduler.client.report [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 667.227389] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "b6e5eefc-8c06-445b-a3af-9404578b6179" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.228377] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "b6e5eefc-8c06-445b-a3af-9404578b6179" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.387049] env[62569]: DEBUG nova.compute.manager [req-3a2aa42c-eb33-45b4-a95e-33aa233c4df0 req-30eb358b-591d-4a42-bc55-2f76bb7f7f3d service nova] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Received event network-changed-a0f99f88-8014-4300-863a-38e81bc640d2 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 667.387049] env[62569]: DEBUG nova.compute.manager [req-3a2aa42c-eb33-45b4-a95e-33aa233c4df0 req-30eb358b-591d-4a42-bc55-2f76bb7f7f3d service nova] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Refreshing instance network info cache due to event network-changed-a0f99f88-8014-4300-863a-38e81bc640d2. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 667.388842] env[62569]: DEBUG oslo_concurrency.lockutils [req-3a2aa42c-eb33-45b4-a95e-33aa233c4df0 req-30eb358b-591d-4a42-bc55-2f76bb7f7f3d service nova] Acquiring lock "refresh_cache-695022e4-bb58-44d7-87e5-58bbd577ec7b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.389148] env[62569]: DEBUG oslo_concurrency.lockutils [req-3a2aa42c-eb33-45b4-a95e-33aa233c4df0 req-30eb358b-591d-4a42-bc55-2f76bb7f7f3d service nova] Acquired lock "refresh_cache-695022e4-bb58-44d7-87e5-58bbd577ec7b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.393016] env[62569]: DEBUG nova.network.neutron [req-3a2aa42c-eb33-45b4-a95e-33aa233c4df0 req-30eb358b-591d-4a42-bc55-2f76bb7f7f3d service nova] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Refreshing network info cache for port a0f99f88-8014-4300-863a-38e81bc640d2 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 667.393016] env[62569]: DEBUG nova.compute.manager [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 667.578500] env[62569]: ERROR nova.compute.manager [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a0f99f88-8014-4300-863a-38e81bc640d2, please check neutron logs for more information. [ 667.578500] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 667.578500] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 667.578500] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 667.578500] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 667.578500] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 667.578500] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 667.578500] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 667.578500] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.578500] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 667.578500] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.578500] env[62569]: ERROR nova.compute.manager raise self.value [ 667.578500] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 667.578500] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 667.578500] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.578500] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 667.579072] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.579072] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 667.579072] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a0f99f88-8014-4300-863a-38e81bc640d2, please check neutron logs for more information. [ 667.579072] env[62569]: ERROR nova.compute.manager [ 667.579072] env[62569]: Traceback (most recent call last): [ 667.579072] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 667.579072] env[62569]: listener.cb(fileno) [ 667.579072] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.579072] env[62569]: result = function(*args, **kwargs) [ 667.579072] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 667.579072] env[62569]: return func(*args, **kwargs) [ 667.579072] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 667.579072] env[62569]: raise e [ 667.579072] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 667.579072] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 667.579072] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 667.579072] env[62569]: created_port_ids = self._update_ports_for_instance( [ 667.579072] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 667.579072] env[62569]: with excutils.save_and_reraise_exception(): [ 667.579072] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.579072] env[62569]: self.force_reraise() [ 667.579072] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.579072] env[62569]: raise self.value [ 667.579072] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 667.579072] env[62569]: updated_port = self._update_port( [ 667.579072] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.579072] env[62569]: _ensure_no_port_binding_failure(port) [ 667.579072] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.579072] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 667.579819] env[62569]: nova.exception.PortBindingFailed: Binding failed for port a0f99f88-8014-4300-863a-38e81bc640d2, please check neutron logs for more information. [ 667.579819] env[62569]: Removing descriptor: 17 [ 667.639528] env[62569]: DEBUG nova.compute.manager [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 667.660299] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.046s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.660957] env[62569]: ERROR nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bc3b6680-2a28-49ee-8a81-8a88f2dbe09f, please check neutron logs for more information. [ 667.660957] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Traceback (most recent call last): [ 667.660957] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 667.660957] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] self.driver.spawn(context, instance, image_meta, [ 667.660957] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 667.660957] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] self._vmops.spawn(context, instance, image_meta, injected_files, [ 667.660957] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 667.660957] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] vm_ref = self.build_virtual_machine(instance, [ 667.660957] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 667.660957] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] vif_infos = vmwarevif.get_vif_info(self._session, [ 667.660957] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 667.661270] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] for vif in network_info: [ 667.661270] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 667.661270] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] return self._sync_wrapper(fn, *args, **kwargs) [ 667.661270] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 667.661270] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] self.wait() [ 667.661270] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 667.661270] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] self[:] = self._gt.wait() [ 667.661270] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 667.661270] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] return self._exit_event.wait() [ 667.661270] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 667.661270] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] current.throw(*self._exc) [ 667.661270] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.661270] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] result = function(*args, **kwargs) [ 667.661620] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 667.661620] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] return func(*args, **kwargs) [ 667.661620] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 667.661620] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] raise e [ 667.661620] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 667.661620] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] nwinfo = self.network_api.allocate_for_instance( [ 667.661620] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 667.661620] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] created_port_ids = self._update_ports_for_instance( [ 667.661620] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 667.661620] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] with excutils.save_and_reraise_exception(): [ 667.661620] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.661620] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] self.force_reraise() [ 667.661620] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.661978] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] raise self.value [ 667.661978] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 667.661978] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] updated_port = self._update_port( [ 667.661978] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.661978] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] _ensure_no_port_binding_failure(port) [ 667.661978] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.661978] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] raise exception.PortBindingFailed(port_id=port['id']) [ 667.661978] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] nova.exception.PortBindingFailed: Binding failed for port bc3b6680-2a28-49ee-8a81-8a88f2dbe09f, please check neutron logs for more information. [ 667.661978] env[62569]: ERROR nova.compute.manager [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] [ 667.661978] env[62569]: DEBUG nova.compute.utils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Binding failed for port bc3b6680-2a28-49ee-8a81-8a88f2dbe09f, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 667.663497] env[62569]: DEBUG nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Build of instance 92666c7b-7a1d-4d49-931f-e84cd2486871 was re-scheduled: Binding failed for port bc3b6680-2a28-49ee-8a81-8a88f2dbe09f, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 667.663718] env[62569]: DEBUG nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 667.663976] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "refresh_cache-92666c7b-7a1d-4d49-931f-e84cd2486871" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.664186] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquired lock "refresh_cache-92666c7b-7a1d-4d49-931f-e84cd2486871" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.664388] env[62569]: DEBUG nova.network.neutron [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 667.667804] env[62569]: DEBUG nova.virt.hardware [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 667.668084] env[62569]: DEBUG nova.virt.hardware [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 667.668281] env[62569]: DEBUG nova.virt.hardware [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 667.668500] env[62569]: DEBUG nova.virt.hardware [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 667.668680] env[62569]: DEBUG nova.virt.hardware [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 667.668859] env[62569]: DEBUG nova.virt.hardware [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 667.669110] env[62569]: DEBUG nova.virt.hardware [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 667.669303] env[62569]: DEBUG nova.virt.hardware [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 667.669504] env[62569]: DEBUG nova.virt.hardware [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 667.669781] env[62569]: DEBUG nova.virt.hardware [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 667.669976] env[62569]: DEBUG nova.virt.hardware [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 667.670343] env[62569]: DEBUG oslo_concurrency.lockutils [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.372s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.673909] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76908d55-efe1-45c4-850a-3a94a2dfac5b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.683473] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce9f5bff-5560-4a07-9b35-09b8c4137829 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.701490] env[62569]: ERROR nova.compute.manager [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a0f99f88-8014-4300-863a-38e81bc640d2, please check neutron logs for more information. [ 667.701490] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Traceback (most recent call last): [ 667.701490] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 667.701490] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] yield resources [ 667.701490] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 667.701490] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] self.driver.spawn(context, instance, image_meta, [ 667.701490] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 667.701490] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 667.701490] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 667.701490] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] vm_ref = self.build_virtual_machine(instance, [ 667.701490] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 667.701919] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] vif_infos = vmwarevif.get_vif_info(self._session, [ 667.701919] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 667.701919] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] for vif in network_info: [ 667.701919] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 667.701919] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] return self._sync_wrapper(fn, *args, **kwargs) [ 667.701919] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 667.701919] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] self.wait() [ 667.701919] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 667.701919] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] self[:] = self._gt.wait() [ 667.701919] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 667.701919] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] return self._exit_event.wait() [ 667.701919] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 667.701919] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] current.throw(*self._exc) [ 667.702308] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.702308] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] result = function(*args, **kwargs) [ 667.702308] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 667.702308] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] return func(*args, **kwargs) [ 667.702308] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 667.702308] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] raise e [ 667.702308] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 667.702308] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] nwinfo = self.network_api.allocate_for_instance( [ 667.702308] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 667.702308] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] created_port_ids = self._update_ports_for_instance( [ 667.702308] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 667.702308] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] with excutils.save_and_reraise_exception(): [ 667.702308] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.703645] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] self.force_reraise() [ 667.703645] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.703645] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] raise self.value [ 667.703645] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 667.703645] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] updated_port = self._update_port( [ 667.703645] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.703645] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] _ensure_no_port_binding_failure(port) [ 667.703645] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.703645] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] raise exception.PortBindingFailed(port_id=port['id']) [ 667.703645] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] nova.exception.PortBindingFailed: Binding failed for port a0f99f88-8014-4300-863a-38e81bc640d2, please check neutron logs for more information. [ 667.703645] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] [ 667.703645] env[62569]: INFO nova.compute.manager [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Terminating instance [ 667.919879] env[62569]: DEBUG nova.network.neutron [req-3a2aa42c-eb33-45b4-a95e-33aa233c4df0 req-30eb358b-591d-4a42-bc55-2f76bb7f7f3d service nova] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.932335] env[62569]: DEBUG oslo_concurrency.lockutils [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.026517] env[62569]: DEBUG nova.network.neutron [req-3a2aa42c-eb33-45b4-a95e-33aa233c4df0 req-30eb358b-591d-4a42-bc55-2f76bb7f7f3d service nova] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.178252] env[62569]: DEBUG nova.compute.utils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Can not refresh info_cache because instance was not found {{(pid=62569) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1024}} [ 668.207787] env[62569]: DEBUG nova.network.neutron [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 668.210633] env[62569]: DEBUG oslo_concurrency.lockutils [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquiring lock "refresh_cache-695022e4-bb58-44d7-87e5-58bbd577ec7b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.284899] env[62569]: DEBUG nova.network.neutron [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.504142] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "37a4c60a-e660-4c91-bcfe-72638a4397b0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.504489] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "37a4c60a-e660-4c91-bcfe-72638a4397b0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.532420] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "06067572-a9fd-43a4-91cd-383a62f53885" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.532643] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "06067572-a9fd-43a4-91cd-383a62f53885" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.532985] env[62569]: DEBUG oslo_concurrency.lockutils [req-3a2aa42c-eb33-45b4-a95e-33aa233c4df0 req-30eb358b-591d-4a42-bc55-2f76bb7f7f3d service nova] Releasing lock "refresh_cache-695022e4-bb58-44d7-87e5-58bbd577ec7b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.533605] env[62569]: DEBUG oslo_concurrency.lockutils [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquired lock "refresh_cache-695022e4-bb58-44d7-87e5-58bbd577ec7b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.534305] env[62569]: DEBUG nova.network.neutron [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 668.616392] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61dc7a91-01e5-47e0-8053-1c9b8b2628c7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.624190] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a6309c-939f-455c-84f0-390ade6e6276 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.655133] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2712e600-4f0b-43d1-81ac-da507ccfe82f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.665972] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65b747e4-e76e-4c65-a02f-2bac2dce2e0d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.678058] env[62569]: DEBUG nova.compute.provider_tree [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.788350] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Releasing lock "refresh_cache-92666c7b-7a1d-4d49-931f-e84cd2486871" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.788632] env[62569]: DEBUG nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 668.788829] env[62569]: DEBUG nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 668.789065] env[62569]: DEBUG nova.network.neutron [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 668.811312] env[62569]: DEBUG nova.network.neutron [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.052104] env[62569]: DEBUG nova.network.neutron [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.131266] env[62569]: DEBUG nova.network.neutron [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.180346] env[62569]: DEBUG nova.scheduler.client.report [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 669.313306] env[62569]: DEBUG nova.network.neutron [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.409882] env[62569]: DEBUG nova.compute.manager [req-44cb98bb-ae3e-43b6-bc1d-3b1ce9e5cf87 req-782a2502-9f1e-476a-b6a4-1c22c421dd08 service nova] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Received event network-vif-deleted-a0f99f88-8014-4300-863a-38e81bc640d2 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 669.633975] env[62569]: DEBUG oslo_concurrency.lockutils [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Releasing lock "refresh_cache-695022e4-bb58-44d7-87e5-58bbd577ec7b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.634411] env[62569]: DEBUG nova.compute.manager [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 669.634603] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 669.634897] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-63518698-7e26-4060-a201-6261627efe28 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.644342] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56fd6b36-410e-431d-8999-549527ce558e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.664692] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 695022e4-bb58-44d7-87e5-58bbd577ec7b could not be found. [ 669.664904] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 669.665097] env[62569]: INFO nova.compute.manager [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 669.665335] env[62569]: DEBUG oslo.service.loopingcall [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 669.665550] env[62569]: DEBUG nova.compute.manager [-] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 669.665662] env[62569]: DEBUG nova.network.neutron [-] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 669.681290] env[62569]: DEBUG nova.network.neutron [-] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.685553] env[62569]: DEBUG oslo_concurrency.lockutils [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.015s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.686163] env[62569]: ERROR nova.compute.manager [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 529ffd30-aece-4729-b8da-d28c3ef597d9, please check neutron logs for more information. [ 669.686163] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Traceback (most recent call last): [ 669.686163] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 669.686163] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] self.driver.spawn(context, instance, image_meta, [ 669.686163] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 669.686163] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 669.686163] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 669.686163] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] vm_ref = self.build_virtual_machine(instance, [ 669.686163] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 669.686163] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] vif_infos = vmwarevif.get_vif_info(self._session, [ 669.686163] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 669.686572] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] for vif in network_info: [ 669.686572] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 669.686572] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] return self._sync_wrapper(fn, *args, **kwargs) [ 669.686572] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 669.686572] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] self.wait() [ 669.686572] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 669.686572] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] self[:] = self._gt.wait() [ 669.686572] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 669.686572] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] return self._exit_event.wait() [ 669.686572] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 669.686572] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] current.throw(*self._exc) [ 669.686572] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.686572] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] result = function(*args, **kwargs) [ 669.686965] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.686965] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] return func(*args, **kwargs) [ 669.686965] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 669.686965] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] raise e [ 669.686965] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 669.686965] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] nwinfo = self.network_api.allocate_for_instance( [ 669.686965] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.686965] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] created_port_ids = self._update_ports_for_instance( [ 669.686965] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.686965] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] with excutils.save_and_reraise_exception(): [ 669.686965] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.686965] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] self.force_reraise() [ 669.686965] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.687365] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] raise self.value [ 669.687365] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.687365] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] updated_port = self._update_port( [ 669.687365] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.687365] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] _ensure_no_port_binding_failure(port) [ 669.687365] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.687365] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] raise exception.PortBindingFailed(port_id=port['id']) [ 669.687365] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] nova.exception.PortBindingFailed: Binding failed for port 529ffd30-aece-4729-b8da-d28c3ef597d9, please check neutron logs for more information. [ 669.687365] env[62569]: ERROR nova.compute.manager [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] [ 669.687365] env[62569]: DEBUG nova.compute.utils [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Binding failed for port 529ffd30-aece-4729-b8da-d28c3ef597d9, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 669.688275] env[62569]: DEBUG oslo_concurrency.lockutils [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.354s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.690506] env[62569]: INFO nova.compute.claims [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 669.694017] env[62569]: DEBUG nova.compute.manager [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Build of instance 21cde60b-2764-4ff7-ab5e-a17fb672351d was re-scheduled: Binding failed for port 529ffd30-aece-4729-b8da-d28c3ef597d9, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 669.694017] env[62569]: DEBUG nova.compute.manager [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 669.694207] env[62569]: DEBUG oslo_concurrency.lockutils [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Acquiring lock "refresh_cache-21cde60b-2764-4ff7-ab5e-a17fb672351d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.694344] env[62569]: DEBUG oslo_concurrency.lockutils [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Acquired lock "refresh_cache-21cde60b-2764-4ff7-ab5e-a17fb672351d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.694497] env[62569]: DEBUG nova.network.neutron [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 669.815545] env[62569]: INFO nova.compute.manager [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 92666c7b-7a1d-4d49-931f-e84cd2486871] Took 1.03 seconds to deallocate network for instance. [ 670.183671] env[62569]: DEBUG nova.network.neutron [-] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.215039] env[62569]: DEBUG nova.network.neutron [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.260517] env[62569]: DEBUG nova.network.neutron [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.687406] env[62569]: INFO nova.compute.manager [-] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Took 1.02 seconds to deallocate network for instance. [ 670.689738] env[62569]: DEBUG nova.compute.claims [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 670.689974] env[62569]: DEBUG oslo_concurrency.lockutils [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.762479] env[62569]: DEBUG oslo_concurrency.lockutils [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Releasing lock "refresh_cache-21cde60b-2764-4ff7-ab5e-a17fb672351d" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.762702] env[62569]: DEBUG nova.compute.manager [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 670.762865] env[62569]: DEBUG nova.compute.manager [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 670.763047] env[62569]: DEBUG nova.network.neutron [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 670.780367] env[62569]: DEBUG nova.network.neutron [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.074619] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6563dbc1-670e-47c3-8a5a-7b8054a7092e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.082659] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4ef5d70-b0cd-4e48-afda-65fdb65eea15 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.112064] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b2e214d-4fd2-43b3-b806-0e72652325f7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.119324] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb1be781-d1a8-4a62-86e0-5b0db3d87dd1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.133040] env[62569]: DEBUG nova.compute.provider_tree [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 671.283344] env[62569]: DEBUG nova.network.neutron [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.635831] env[62569]: DEBUG nova.scheduler.client.report [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 671.786683] env[62569]: INFO nova.compute.manager [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] [instance: 21cde60b-2764-4ff7-ab5e-a17fb672351d] Took 1.02 seconds to deallocate network for instance. [ 671.849310] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f50dd663-bd97-476c-9900-edfe03013654 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "92666c7b-7a1d-4d49-931f-e84cd2486871" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.684s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.849938] env[62569]: Traceback (most recent call last): [ 671.849992] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 671.849992] env[62569]: self.driver.spawn(context, instance, image_meta, [ 671.849992] env[62569]: File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 671.849992] env[62569]: self._vmops.spawn(context, instance, image_meta, injected_files, [ 671.849992] env[62569]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 671.849992] env[62569]: vm_ref = self.build_virtual_machine(instance, [ 671.849992] env[62569]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 671.849992] env[62569]: vif_infos = vmwarevif.get_vif_info(self._session, [ 671.849992] env[62569]: File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 671.849992] env[62569]: for vif in network_info: [ 671.849992] env[62569]: File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 671.849992] env[62569]: return self._sync_wrapper(fn, *args, **kwargs) [ 671.849992] env[62569]: File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 671.849992] env[62569]: self.wait() [ 671.849992] env[62569]: File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 671.849992] env[62569]: self[:] = self._gt.wait() [ 671.849992] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 671.849992] env[62569]: return self._exit_event.wait() [ 671.850550] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 671.850550] env[62569]: current.throw(*self._exc) [ 671.850550] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 671.850550] env[62569]: result = function(*args, **kwargs) [ 671.850550] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 671.850550] env[62569]: return func(*args, **kwargs) [ 671.850550] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 671.850550] env[62569]: raise e [ 671.850550] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 671.850550] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 671.850550] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 671.850550] env[62569]: created_port_ids = self._update_ports_for_instance( [ 671.850550] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 671.850550] env[62569]: with excutils.save_and_reraise_exception(): [ 671.850550] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.850550] env[62569]: self.force_reraise() [ 671.850550] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.850550] env[62569]: raise self.value [ 671.850550] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 671.850550] env[62569]: updated_port = self._update_port( [ 671.850550] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.850550] env[62569]: _ensure_no_port_binding_failure(port) [ 671.850550] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.850550] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 671.850550] env[62569]: nova.exception.PortBindingFailed: Binding failed for port bc3b6680-2a28-49ee-8a81-8a88f2dbe09f, please check neutron logs for more information. [ 671.850550] env[62569]: During handling of the above exception, another exception occurred: [ 671.850550] env[62569]: Traceback (most recent call last): [ 671.850550] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2462, in _do_build_and_run_instance [ 671.850550] env[62569]: self._build_and_run_instance(context, instance, image, [ 671.850550] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2754, in _build_and_run_instance [ 671.851371] env[62569]: raise exception.RescheduledException( [ 671.851371] env[62569]: nova.exception.RescheduledException: Build of instance 92666c7b-7a1d-4d49-931f-e84cd2486871 was re-scheduled: Binding failed for port bc3b6680-2a28-49ee-8a81-8a88f2dbe09f, please check neutron logs for more information. [ 671.851371] env[62569]: During handling of the above exception, another exception occurred: [ 671.851371] env[62569]: Traceback (most recent call last): [ 671.851371] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 671.851371] env[62569]: func(*args, **kwargs) [ 671.851371] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 671.851371] env[62569]: return func(*args, **kwargs) [ 671.851371] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 671.851371] env[62569]: return f(*args, **kwargs) [ 671.851371] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2353, in _locked_do_build_and_run_instance [ 671.851371] env[62569]: result = self._do_build_and_run_instance(*args, **kwargs) [ 671.851371] env[62569]: File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 671.851371] env[62569]: with excutils.save_and_reraise_exception(): [ 671.851371] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.851371] env[62569]: self.force_reraise() [ 671.851371] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.851371] env[62569]: raise self.value [ 671.851371] env[62569]: File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 671.851371] env[62569]: return f(self, context, *args, **kw) [ 671.851371] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 168, in decorated_function [ 671.851371] env[62569]: with excutils.save_and_reraise_exception(): [ 671.851371] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.851371] env[62569]: self.force_reraise() [ 671.851371] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.851371] env[62569]: raise self.value [ 671.851371] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 159, in decorated_function [ 671.851371] env[62569]: return function(self, context, *args, **kwargs) [ 671.851371] env[62569]: File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 671.851371] env[62569]: return function(self, context, *args, **kwargs) [ 671.852259] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 205, in decorated_function [ 671.852259] env[62569]: return function(self, context, *args, **kwargs) [ 671.852259] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2497, in _do_build_and_run_instance [ 671.852259] env[62569]: instance.save() [ 671.852259] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 209, in wrapper [ 671.852259] env[62569]: updates, result = self.indirection_api.object_action( [ 671.852259] env[62569]: File "/opt/stack/nova/nova/conductor/rpcapi.py", line 247, in object_action [ 671.852259] env[62569]: return cctxt.call(context, 'object_action', objinst=objinst, [ 671.852259] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/client.py", line 190, in call [ 671.852259] env[62569]: result = self.transport._send( [ 671.852259] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/transport.py", line 123, in _send [ 671.852259] env[62569]: return self._driver.send(target, ctxt, message, [ 671.852259] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 788, in send [ 671.852259] env[62569]: return self._send(target, ctxt, message, wait_for_reply, timeout, [ 671.852259] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 780, in _send [ 671.852259] env[62569]: raise result [ 671.852259] env[62569]: nova.exception_Remote.InstanceNotFound_Remote: Instance 92666c7b-7a1d-4d49-931f-e84cd2486871 could not be found. [ 671.852259] env[62569]: Traceback (most recent call last): [ 671.852259] env[62569]: File "/opt/stack/nova/nova/conductor/manager.py", line 142, in _object_dispatch [ 671.852259] env[62569]: return getattr(target, method)(*args, **kwargs) [ 671.852259] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 226, in wrapper [ 671.852259] env[62569]: return fn(self, *args, **kwargs) [ 671.852259] env[62569]: File "/opt/stack/nova/nova/objects/instance.py", line 878, in save [ 671.852259] env[62569]: old_ref, inst_ref = db.instance_update_and_get_original( [ 671.852259] env[62569]: File "/opt/stack/nova/nova/db/utils.py", line 35, in wrapper [ 671.852259] env[62569]: return f(*args, **kwargs) [ 671.852259] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 144, in wrapper [ 671.852259] env[62569]: with excutils.save_and_reraise_exception() as ectxt: [ 671.853105] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.853105] env[62569]: self.force_reraise() [ 671.853105] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.853105] env[62569]: raise self.value [ 671.853105] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 142, in wrapper [ 671.853105] env[62569]: return f(*args, **kwargs) [ 671.853105] env[62569]: File "/opt/stack/nova/nova/db/main/api.py", line 207, in wrapper [ 671.853105] env[62569]: return f(context, *args, **kwargs) [ 671.853105] env[62569]: File "/opt/stack/nova/nova/db/main/api.py", line 2283, in instance_update_and_get_original [ 671.853105] env[62569]: instance_ref = _instance_get_by_uuid(context, instance_uuid, [ 671.853105] env[62569]: File "/opt/stack/nova/nova/db/main/api.py", line 1405, in _instance_get_by_uuid [ 671.853105] env[62569]: raise exception.InstanceNotFound(instance_id=uuid) [ 671.853105] env[62569]: nova.exception.InstanceNotFound: Instance 92666c7b-7a1d-4d49-931f-e84cd2486871 could not be found. [ 672.142850] env[62569]: DEBUG oslo_concurrency.lockutils [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.455s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.143421] env[62569]: DEBUG nova.compute.manager [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 672.145937] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.018s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.353106] env[62569]: DEBUG nova.compute.manager [None req-cacfbb43-9283-453e-9044-851bf0c24b47 tempest-ServersListShow296Test-1909070751 tempest-ServersListShow296Test-1909070751-project-member] [instance: d0fbb20f-bfcc-491e-a7ec-e7e04a0d11fe] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 672.649942] env[62569]: DEBUG nova.compute.utils [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 672.654544] env[62569]: DEBUG nova.compute.manager [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 672.654658] env[62569]: DEBUG nova.network.neutron [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 672.701567] env[62569]: DEBUG nova.policy [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1418d7ccc2a4324aeb586f89c2f817b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4ba7e96d226942bd99ab29fe703c6fea', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 672.815527] env[62569]: INFO nova.scheduler.client.report [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Deleted allocations for instance 21cde60b-2764-4ff7-ab5e-a17fb672351d [ 672.857450] env[62569]: DEBUG nova.compute.manager [None req-cacfbb43-9283-453e-9044-851bf0c24b47 tempest-ServersListShow296Test-1909070751 tempest-ServersListShow296Test-1909070751-project-member] [instance: d0fbb20f-bfcc-491e-a7ec-e7e04a0d11fe] Instance disappeared before build. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2445}} [ 673.007118] env[62569]: DEBUG nova.network.neutron [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Successfully created port: e4af3f09-3554-4008-bd79-56dd4a9affe7 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 673.055713] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b087caba-5f1b-462a-8569-7176bea53ce9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.063585] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a162957e-bada-413e-bc7b-ee1b84a445fb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.096199] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a4c27eb-7a9b-4f5d-9dd6-8da45da3f640 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.102053] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1694db89-2645-4f74-91a8-bcb96d9763d6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.115093] env[62569]: DEBUG nova.compute.provider_tree [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 673.157677] env[62569]: DEBUG nova.compute.manager [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 673.327359] env[62569]: DEBUG oslo_concurrency.lockutils [None req-312492dd-c99d-496d-94f5-cd76c5253a3a tempest-ServerDiagnosticsNegativeTest-1567500594 tempest-ServerDiagnosticsNegativeTest-1567500594-project-member] Lock "21cde60b-2764-4ff7-ab5e-a17fb672351d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.650s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.377742] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cacfbb43-9283-453e-9044-851bf0c24b47 tempest-ServersListShow296Test-1909070751 tempest-ServersListShow296Test-1909070751-project-member] Lock "d0fbb20f-bfcc-491e-a7ec-e7e04a0d11fe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.714s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.618442] env[62569]: DEBUG nova.scheduler.client.report [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 673.782158] env[62569]: DEBUG nova.compute.manager [req-cdd31d16-577c-4ffa-806d-0099645214eb req-51033ce6-5c76-4d57-8e83-9595ff35d8eb service nova] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Received event network-changed-e4af3f09-3554-4008-bd79-56dd4a9affe7 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 673.782158] env[62569]: DEBUG nova.compute.manager [req-cdd31d16-577c-4ffa-806d-0099645214eb req-51033ce6-5c76-4d57-8e83-9595ff35d8eb service nova] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Refreshing instance network info cache due to event network-changed-e4af3f09-3554-4008-bd79-56dd4a9affe7. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 673.782158] env[62569]: DEBUG oslo_concurrency.lockutils [req-cdd31d16-577c-4ffa-806d-0099645214eb req-51033ce6-5c76-4d57-8e83-9595ff35d8eb service nova] Acquiring lock "refresh_cache-0eff7f21-17f8-4642-bfc4-a7a15bdfb41f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.782158] env[62569]: DEBUG oslo_concurrency.lockutils [req-cdd31d16-577c-4ffa-806d-0099645214eb req-51033ce6-5c76-4d57-8e83-9595ff35d8eb service nova] Acquired lock "refresh_cache-0eff7f21-17f8-4642-bfc4-a7a15bdfb41f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.782158] env[62569]: DEBUG nova.network.neutron [req-cdd31d16-577c-4ffa-806d-0099645214eb req-51033ce6-5c76-4d57-8e83-9595ff35d8eb service nova] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Refreshing network info cache for port e4af3f09-3554-4008-bd79-56dd4a9affe7 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 673.828808] env[62569]: DEBUG nova.compute.manager [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 673.882859] env[62569]: DEBUG nova.compute.manager [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 673.988643] env[62569]: ERROR nova.compute.manager [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e4af3f09-3554-4008-bd79-56dd4a9affe7, please check neutron logs for more information. [ 673.988643] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 673.988643] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 673.988643] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 673.988643] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 673.988643] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 673.988643] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 673.988643] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 673.988643] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 673.988643] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 673.988643] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 673.988643] env[62569]: ERROR nova.compute.manager raise self.value [ 673.988643] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 673.988643] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 673.988643] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 673.988643] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 673.989185] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 673.989185] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 673.989185] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e4af3f09-3554-4008-bd79-56dd4a9affe7, please check neutron logs for more information. [ 673.989185] env[62569]: ERROR nova.compute.manager [ 673.989686] env[62569]: Traceback (most recent call last): [ 673.989686] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 673.989686] env[62569]: listener.cb(fileno) [ 673.989686] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 673.989686] env[62569]: result = function(*args, **kwargs) [ 673.989686] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 673.989686] env[62569]: return func(*args, **kwargs) [ 673.989686] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 673.989686] env[62569]: raise e [ 673.989686] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 673.989686] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 673.989686] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 673.989686] env[62569]: created_port_ids = self._update_ports_for_instance( [ 673.989686] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 673.989686] env[62569]: with excutils.save_and_reraise_exception(): [ 673.989686] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 673.989686] env[62569]: self.force_reraise() [ 673.989686] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 673.989686] env[62569]: raise self.value [ 673.989686] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 673.989686] env[62569]: updated_port = self._update_port( [ 673.989686] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 673.989686] env[62569]: _ensure_no_port_binding_failure(port) [ 673.989686] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 673.989686] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 673.989686] env[62569]: nova.exception.PortBindingFailed: Binding failed for port e4af3f09-3554-4008-bd79-56dd4a9affe7, please check neutron logs for more information. [ 673.989686] env[62569]: Removing descriptor: 17 [ 674.126784] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.981s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.128074] env[62569]: ERROR nova.compute.manager [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d3caf564-b121-4a3b-928a-6b8bc449265b, please check neutron logs for more information. [ 674.128074] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Traceback (most recent call last): [ 674.128074] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 674.128074] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] self.driver.spawn(context, instance, image_meta, [ 674.128074] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 674.128074] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] self._vmops.spawn(context, instance, image_meta, injected_files, [ 674.128074] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 674.128074] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] vm_ref = self.build_virtual_machine(instance, [ 674.128074] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 674.128074] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] vif_infos = vmwarevif.get_vif_info(self._session, [ 674.128074] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 674.129907] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] for vif in network_info: [ 674.129907] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 674.129907] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] return self._sync_wrapper(fn, *args, **kwargs) [ 674.129907] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 674.129907] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] self.wait() [ 674.129907] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 674.129907] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] self[:] = self._gt.wait() [ 674.129907] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 674.129907] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] return self._exit_event.wait() [ 674.129907] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 674.129907] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] current.throw(*self._exc) [ 674.129907] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.129907] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] result = function(*args, **kwargs) [ 674.130304] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 674.130304] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] return func(*args, **kwargs) [ 674.130304] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 674.130304] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] raise e [ 674.130304] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 674.130304] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] nwinfo = self.network_api.allocate_for_instance( [ 674.130304] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 674.130304] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] created_port_ids = self._update_ports_for_instance( [ 674.130304] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 674.130304] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] with excutils.save_and_reraise_exception(): [ 674.130304] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.130304] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] self.force_reraise() [ 674.130304] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.130688] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] raise self.value [ 674.130688] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 674.130688] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] updated_port = self._update_port( [ 674.130688] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.130688] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] _ensure_no_port_binding_failure(port) [ 674.130688] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.130688] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] raise exception.PortBindingFailed(port_id=port['id']) [ 674.130688] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] nova.exception.PortBindingFailed: Binding failed for port d3caf564-b121-4a3b-928a-6b8bc449265b, please check neutron logs for more information. [ 674.130688] env[62569]: ERROR nova.compute.manager [instance: d53a6920-d70c-40a6-a059-162a5f876b29] [ 674.130688] env[62569]: DEBUG nova.compute.utils [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Binding failed for port d3caf564-b121-4a3b-928a-6b8bc449265b, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 674.131053] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.696s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.133282] env[62569]: DEBUG nova.compute.manager [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Build of instance d53a6920-d70c-40a6-a059-162a5f876b29 was re-scheduled: Binding failed for port d3caf564-b121-4a3b-928a-6b8bc449265b, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 674.133765] env[62569]: DEBUG nova.compute.manager [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 674.134032] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Acquiring lock "refresh_cache-d53a6920-d70c-40a6-a059-162a5f876b29" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.134214] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Acquired lock "refresh_cache-d53a6920-d70c-40a6-a059-162a5f876b29" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.134529] env[62569]: DEBUG nova.network.neutron [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 674.164632] env[62569]: DEBUG nova.compute.manager [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 674.189444] env[62569]: DEBUG nova.virt.hardware [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 674.189685] env[62569]: DEBUG nova.virt.hardware [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 674.189901] env[62569]: DEBUG nova.virt.hardware [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 674.190182] env[62569]: DEBUG nova.virt.hardware [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 674.190339] env[62569]: DEBUG nova.virt.hardware [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 674.190491] env[62569]: DEBUG nova.virt.hardware [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 674.190695] env[62569]: DEBUG nova.virt.hardware [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 674.190981] env[62569]: DEBUG nova.virt.hardware [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 674.191061] env[62569]: DEBUG nova.virt.hardware [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 674.191191] env[62569]: DEBUG nova.virt.hardware [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 674.191361] env[62569]: DEBUG nova.virt.hardware [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 674.192244] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77304878-bb77-4004-8402-04aca5f0d4c5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.200714] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daf6b7d5-44dd-455f-839e-ace9dc56c437 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.215059] env[62569]: ERROR nova.compute.manager [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e4af3f09-3554-4008-bd79-56dd4a9affe7, please check neutron logs for more information. [ 674.215059] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Traceback (most recent call last): [ 674.215059] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 674.215059] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] yield resources [ 674.215059] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 674.215059] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] self.driver.spawn(context, instance, image_meta, [ 674.215059] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 674.215059] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 674.215059] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 674.215059] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] vm_ref = self.build_virtual_machine(instance, [ 674.215059] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 674.215406] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] vif_infos = vmwarevif.get_vif_info(self._session, [ 674.215406] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 674.215406] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] for vif in network_info: [ 674.215406] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 674.215406] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] return self._sync_wrapper(fn, *args, **kwargs) [ 674.215406] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 674.215406] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] self.wait() [ 674.215406] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 674.215406] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] self[:] = self._gt.wait() [ 674.215406] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 674.215406] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] return self._exit_event.wait() [ 674.215406] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 674.215406] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] current.throw(*self._exc) [ 674.215790] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.215790] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] result = function(*args, **kwargs) [ 674.215790] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 674.215790] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] return func(*args, **kwargs) [ 674.215790] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 674.215790] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] raise e [ 674.215790] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 674.215790] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] nwinfo = self.network_api.allocate_for_instance( [ 674.215790] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 674.215790] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] created_port_ids = self._update_ports_for_instance( [ 674.215790] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 674.215790] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] with excutils.save_and_reraise_exception(): [ 674.215790] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.216178] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] self.force_reraise() [ 674.216178] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.216178] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] raise self.value [ 674.216178] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 674.216178] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] updated_port = self._update_port( [ 674.216178] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.216178] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] _ensure_no_port_binding_failure(port) [ 674.216178] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.216178] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] raise exception.PortBindingFailed(port_id=port['id']) [ 674.216178] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] nova.exception.PortBindingFailed: Binding failed for port e4af3f09-3554-4008-bd79-56dd4a9affe7, please check neutron logs for more information. [ 674.216178] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] [ 674.216178] env[62569]: INFO nova.compute.manager [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Terminating instance [ 674.299915] env[62569]: DEBUG nova.network.neutron [req-cdd31d16-577c-4ffa-806d-0099645214eb req-51033ce6-5c76-4d57-8e83-9595ff35d8eb service nova] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.356157] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.383930] env[62569]: DEBUG nova.network.neutron [req-cdd31d16-577c-4ffa-806d-0099645214eb req-51033ce6-5c76-4d57-8e83-9595ff35d8eb service nova] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.403451] env[62569]: DEBUG oslo_concurrency.lockutils [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.658161] env[62569]: DEBUG nova.network.neutron [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.720341] env[62569]: DEBUG oslo_concurrency.lockutils [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "refresh_cache-0eff7f21-17f8-4642-bfc4-a7a15bdfb41f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.757422] env[62569]: DEBUG nova.network.neutron [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.888290] env[62569]: DEBUG oslo_concurrency.lockutils [req-cdd31d16-577c-4ffa-806d-0099645214eb req-51033ce6-5c76-4d57-8e83-9595ff35d8eb service nova] Releasing lock "refresh_cache-0eff7f21-17f8-4642-bfc4-a7a15bdfb41f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.888697] env[62569]: DEBUG oslo_concurrency.lockutils [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired lock "refresh_cache-0eff7f21-17f8-4642-bfc4-a7a15bdfb41f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.888988] env[62569]: DEBUG nova.network.neutron [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 675.017447] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6774b746-28fa-426f-89c1-b1a8d5d6baa9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.025075] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aabdd014-6df0-45c6-ba8b-3d30f4512ae5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.055154] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2825adf0-7c25-448a-88e3-e765a56d7a8a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.062079] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0a05af1-8b0a-40b1-9134-f2ef9fba505e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.074734] env[62569]: DEBUG nova.compute.provider_tree [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 675.262150] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Releasing lock "refresh_cache-d53a6920-d70c-40a6-a059-162a5f876b29" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.262406] env[62569]: DEBUG nova.compute.manager [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 675.262594] env[62569]: DEBUG nova.compute.manager [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 675.262764] env[62569]: DEBUG nova.network.neutron [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 675.278490] env[62569]: DEBUG nova.network.neutron [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 675.407219] env[62569]: DEBUG nova.network.neutron [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 675.489554] env[62569]: DEBUG nova.network.neutron [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.578383] env[62569]: DEBUG nova.scheduler.client.report [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 675.780855] env[62569]: DEBUG nova.network.neutron [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.807207] env[62569]: DEBUG nova.compute.manager [req-0ace631f-3385-456e-b739-e7116172e3fb req-18adaa3f-75da-4c2e-bc1b-8a661a17c6a0 service nova] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Received event network-vif-deleted-e4af3f09-3554-4008-bd79-56dd4a9affe7 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 675.992530] env[62569]: DEBUG oslo_concurrency.lockutils [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Releasing lock "refresh_cache-0eff7f21-17f8-4642-bfc4-a7a15bdfb41f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.993026] env[62569]: DEBUG nova.compute.manager [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 675.993228] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 675.993540] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-35dbd701-9c3c-464c-aa82-176664fb1fdc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.002288] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-703219d7-b9cb-4595-91e7-7e4839bc99b7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.022881] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f could not be found. [ 676.023109] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 676.023292] env[62569]: INFO nova.compute.manager [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Took 0.03 seconds to destroy the instance on the hypervisor. [ 676.023528] env[62569]: DEBUG oslo.service.loopingcall [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 676.023729] env[62569]: DEBUG nova.compute.manager [-] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 676.023822] env[62569]: DEBUG nova.network.neutron [-] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 676.039143] env[62569]: DEBUG nova.network.neutron [-] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.082865] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.953s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.083484] env[62569]: ERROR nova.compute.manager [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c03d22b9-2146-448a-aae4-6bb8f97e497e, please check neutron logs for more information. [ 676.083484] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] Traceback (most recent call last): [ 676.083484] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 676.083484] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] self.driver.spawn(context, instance, image_meta, [ 676.083484] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 676.083484] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] self._vmops.spawn(context, instance, image_meta, injected_files, [ 676.083484] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 676.083484] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] vm_ref = self.build_virtual_machine(instance, [ 676.083484] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 676.083484] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] vif_infos = vmwarevif.get_vif_info(self._session, [ 676.083484] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 676.083791] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] for vif in network_info: [ 676.083791] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 676.083791] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] return self._sync_wrapper(fn, *args, **kwargs) [ 676.083791] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 676.083791] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] self.wait() [ 676.083791] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 676.083791] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] self[:] = self._gt.wait() [ 676.083791] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 676.083791] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] return self._exit_event.wait() [ 676.083791] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 676.083791] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] current.throw(*self._exc) [ 676.083791] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 676.083791] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] result = function(*args, **kwargs) [ 676.084113] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 676.084113] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] return func(*args, **kwargs) [ 676.084113] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 676.084113] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] raise e [ 676.084113] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 676.084113] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] nwinfo = self.network_api.allocate_for_instance( [ 676.084113] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 676.084113] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] created_port_ids = self._update_ports_for_instance( [ 676.084113] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 676.084113] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] with excutils.save_and_reraise_exception(): [ 676.084113] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.084113] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] self.force_reraise() [ 676.084113] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.084432] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] raise self.value [ 676.084432] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 676.084432] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] updated_port = self._update_port( [ 676.084432] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.084432] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] _ensure_no_port_binding_failure(port) [ 676.084432] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.084432] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] raise exception.PortBindingFailed(port_id=port['id']) [ 676.084432] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] nova.exception.PortBindingFailed: Binding failed for port c03d22b9-2146-448a-aae4-6bb8f97e497e, please check neutron logs for more information. [ 676.084432] env[62569]: ERROR nova.compute.manager [instance: e53104ad-35c7-49cb-b8de-035006148c52] [ 676.084432] env[62569]: DEBUG nova.compute.utils [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Binding failed for port c03d22b9-2146-448a-aae4-6bb8f97e497e, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 676.085683] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.642s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.087173] env[62569]: INFO nova.compute.claims [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 676.089780] env[62569]: DEBUG nova.compute.manager [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Build of instance e53104ad-35c7-49cb-b8de-035006148c52 was re-scheduled: Binding failed for port c03d22b9-2146-448a-aae4-6bb8f97e497e, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 676.090275] env[62569]: DEBUG nova.compute.manager [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 676.090498] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Acquiring lock "refresh_cache-e53104ad-35c7-49cb-b8de-035006148c52" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.090681] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Acquired lock "refresh_cache-e53104ad-35c7-49cb-b8de-035006148c52" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.090944] env[62569]: DEBUG nova.network.neutron [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 676.283773] env[62569]: INFO nova.compute.manager [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] [instance: d53a6920-d70c-40a6-a059-162a5f876b29] Took 1.02 seconds to deallocate network for instance. [ 676.541754] env[62569]: DEBUG nova.network.neutron [-] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.610522] env[62569]: DEBUG nova.network.neutron [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.685443] env[62569]: DEBUG nova.network.neutron [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.045064] env[62569]: INFO nova.compute.manager [-] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Took 1.02 seconds to deallocate network for instance. [ 677.047265] env[62569]: DEBUG nova.compute.claims [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 677.047421] env[62569]: DEBUG oslo_concurrency.lockutils [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.188208] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Releasing lock "refresh_cache-e53104ad-35c7-49cb-b8de-035006148c52" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.189531] env[62569]: DEBUG nova.compute.manager [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 677.189531] env[62569]: DEBUG nova.compute.manager [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 677.189531] env[62569]: DEBUG nova.network.neutron [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 677.203406] env[62569]: DEBUG nova.network.neutron [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.311403] env[62569]: INFO nova.scheduler.client.report [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Deleted allocations for instance d53a6920-d70c-40a6-a059-162a5f876b29 [ 677.449084] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be8f0968-2c1b-4aa0-8ff2-ac4e6255605b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.456469] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a470cb42-d183-4cd1-b658-c0d8ffd0e33d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.486319] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9de0a78-dc11-4bf3-951b-077f35f6959b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.493397] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb0dfa9a-1bda-41f2-bb24-a28c7e4921e7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.507046] env[62569]: DEBUG nova.compute.provider_tree [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 677.709135] env[62569]: DEBUG nova.network.neutron [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.822652] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fa295451-86a8-4837-839c-8516eab11471 tempest-ImagesNegativeTestJSON-1643185633 tempest-ImagesNegativeTestJSON-1643185633-project-member] Lock "d53a6920-d70c-40a6-a059-162a5f876b29" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.865s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.010748] env[62569]: DEBUG nova.scheduler.client.report [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 678.212351] env[62569]: INFO nova.compute.manager [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] [instance: e53104ad-35c7-49cb-b8de-035006148c52] Took 1.02 seconds to deallocate network for instance. [ 678.325943] env[62569]: DEBUG nova.compute.manager [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 678.515192] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.515724] env[62569]: DEBUG nova.compute.manager [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 678.524366] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.146s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.526584] env[62569]: INFO nova.compute.claims [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 678.855104] env[62569]: DEBUG oslo_concurrency.lockutils [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.026218] env[62569]: DEBUG nova.compute.utils [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 679.029206] env[62569]: DEBUG nova.compute.manager [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 679.029206] env[62569]: DEBUG nova.network.neutron [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 679.072321] env[62569]: DEBUG nova.policy [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de926d5954a148d0be18f305c53e4495', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dbde8d8f33cc4cb5a05c0f94f872d6c2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 679.250156] env[62569]: INFO nova.scheduler.client.report [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Deleted allocations for instance e53104ad-35c7-49cb-b8de-035006148c52 [ 679.370924] env[62569]: DEBUG nova.network.neutron [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Successfully created port: a9cadb21-f24f-480a-819a-10cfdb8e514b {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 679.531798] env[62569]: DEBUG nova.compute.manager [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 679.763940] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47147089-8306-42e7-9984-d95274e48341 tempest-AttachInterfacesUnderV243Test-1920533412 tempest-AttachInterfacesUnderV243Test-1920533412-project-member] Lock "e53104ad-35c7-49cb-b8de-035006148c52" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.840s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.006734] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69f902fb-6343-4f17-bbce-249e6c57a996 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.018232] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac1e0805-776b-4385-8f79-82e40d97511c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.062252] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d0f1e1f-23ae-407c-8932-b7e53576ad51 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.067812] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91371934-7a49-44ed-8393-e9293a77b454 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.083423] env[62569]: DEBUG nova.compute.provider_tree [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 680.214246] env[62569]: DEBUG nova.compute.manager [req-ad9acd59-10f7-4789-b147-fc682867d940 req-8b4bf451-3615-467c-baeb-b8d4f2ce1f41 service nova] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Received event network-changed-a9cadb21-f24f-480a-819a-10cfdb8e514b {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 680.214481] env[62569]: DEBUG nova.compute.manager [req-ad9acd59-10f7-4789-b147-fc682867d940 req-8b4bf451-3615-467c-baeb-b8d4f2ce1f41 service nova] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Refreshing instance network info cache due to event network-changed-a9cadb21-f24f-480a-819a-10cfdb8e514b. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 680.214784] env[62569]: DEBUG oslo_concurrency.lockutils [req-ad9acd59-10f7-4789-b147-fc682867d940 req-8b4bf451-3615-467c-baeb-b8d4f2ce1f41 service nova] Acquiring lock "refresh_cache-0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.214952] env[62569]: DEBUG oslo_concurrency.lockutils [req-ad9acd59-10f7-4789-b147-fc682867d940 req-8b4bf451-3615-467c-baeb-b8d4f2ce1f41 service nova] Acquired lock "refresh_cache-0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.215223] env[62569]: DEBUG nova.network.neutron [req-ad9acd59-10f7-4789-b147-fc682867d940 req-8b4bf451-3615-467c-baeb-b8d4f2ce1f41 service nova] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Refreshing network info cache for port a9cadb21-f24f-480a-819a-10cfdb8e514b {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 680.269684] env[62569]: DEBUG nova.compute.manager [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 680.315120] env[62569]: ERROR nova.compute.manager [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a9cadb21-f24f-480a-819a-10cfdb8e514b, please check neutron logs for more information. [ 680.315120] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 680.315120] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 680.315120] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 680.315120] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.315120] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 680.315120] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.315120] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 680.315120] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.315120] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 680.315120] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.315120] env[62569]: ERROR nova.compute.manager raise self.value [ 680.315120] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.315120] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 680.315120] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.315120] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 680.315632] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.315632] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 680.315632] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a9cadb21-f24f-480a-819a-10cfdb8e514b, please check neutron logs for more information. [ 680.315632] env[62569]: ERROR nova.compute.manager [ 680.315632] env[62569]: Traceback (most recent call last): [ 680.315632] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 680.315632] env[62569]: listener.cb(fileno) [ 680.315632] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.315632] env[62569]: result = function(*args, **kwargs) [ 680.315632] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 680.315632] env[62569]: return func(*args, **kwargs) [ 680.315632] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 680.315632] env[62569]: raise e [ 680.315632] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 680.315632] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 680.315632] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.315632] env[62569]: created_port_ids = self._update_ports_for_instance( [ 680.315632] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.315632] env[62569]: with excutils.save_and_reraise_exception(): [ 680.315632] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.315632] env[62569]: self.force_reraise() [ 680.315632] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.315632] env[62569]: raise self.value [ 680.315632] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.315632] env[62569]: updated_port = self._update_port( [ 680.315632] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.315632] env[62569]: _ensure_no_port_binding_failure(port) [ 680.315632] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.315632] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 680.316520] env[62569]: nova.exception.PortBindingFailed: Binding failed for port a9cadb21-f24f-480a-819a-10cfdb8e514b, please check neutron logs for more information. [ 680.316520] env[62569]: Removing descriptor: 17 [ 680.561287] env[62569]: DEBUG nova.compute.manager [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 680.586188] env[62569]: DEBUG nova.scheduler.client.report [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 680.591494] env[62569]: DEBUG nova.virt.hardware [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 680.591728] env[62569]: DEBUG nova.virt.hardware [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 680.591883] env[62569]: DEBUG nova.virt.hardware [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 680.592073] env[62569]: DEBUG nova.virt.hardware [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 680.592262] env[62569]: DEBUG nova.virt.hardware [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 680.592404] env[62569]: DEBUG nova.virt.hardware [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 680.592605] env[62569]: DEBUG nova.virt.hardware [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 680.592762] env[62569]: DEBUG nova.virt.hardware [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 680.592927] env[62569]: DEBUG nova.virt.hardware [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 680.593344] env[62569]: DEBUG nova.virt.hardware [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 680.593619] env[62569]: DEBUG nova.virt.hardware [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 680.594937] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51f5bf09-4f7b-46af-ab6e-46d610e4dfa4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.606268] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dc0ed3a-8385-4abf-a234-3770532cb520 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.623988] env[62569]: ERROR nova.compute.manager [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a9cadb21-f24f-480a-819a-10cfdb8e514b, please check neutron logs for more information. [ 680.623988] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Traceback (most recent call last): [ 680.623988] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 680.623988] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] yield resources [ 680.623988] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 680.623988] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] self.driver.spawn(context, instance, image_meta, [ 680.623988] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 680.623988] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] self._vmops.spawn(context, instance, image_meta, injected_files, [ 680.623988] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 680.623988] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] vm_ref = self.build_virtual_machine(instance, [ 680.623988] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 680.624365] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] vif_infos = vmwarevif.get_vif_info(self._session, [ 680.624365] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 680.624365] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] for vif in network_info: [ 680.624365] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 680.624365] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] return self._sync_wrapper(fn, *args, **kwargs) [ 680.624365] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 680.624365] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] self.wait() [ 680.624365] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 680.624365] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] self[:] = self._gt.wait() [ 680.624365] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 680.624365] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] return self._exit_event.wait() [ 680.624365] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 680.624365] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] current.throw(*self._exc) [ 680.624719] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.624719] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] result = function(*args, **kwargs) [ 680.624719] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 680.624719] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] return func(*args, **kwargs) [ 680.624719] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 680.624719] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] raise e [ 680.624719] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 680.624719] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] nwinfo = self.network_api.allocate_for_instance( [ 680.624719] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.624719] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] created_port_ids = self._update_ports_for_instance( [ 680.624719] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.624719] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] with excutils.save_and_reraise_exception(): [ 680.624719] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.625078] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] self.force_reraise() [ 680.625078] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.625078] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] raise self.value [ 680.625078] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.625078] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] updated_port = self._update_port( [ 680.625078] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.625078] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] _ensure_no_port_binding_failure(port) [ 680.625078] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.625078] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] raise exception.PortBindingFailed(port_id=port['id']) [ 680.625078] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] nova.exception.PortBindingFailed: Binding failed for port a9cadb21-f24f-480a-819a-10cfdb8e514b, please check neutron logs for more information. [ 680.625078] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] [ 680.625078] env[62569]: INFO nova.compute.manager [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Terminating instance [ 680.737314] env[62569]: DEBUG nova.network.neutron [req-ad9acd59-10f7-4789-b147-fc682867d940 req-8b4bf451-3615-467c-baeb-b8d4f2ce1f41 service nova] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.792122] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.864362] env[62569]: DEBUG nova.network.neutron [req-ad9acd59-10f7-4789-b147-fc682867d940 req-8b4bf451-3615-467c-baeb-b8d4f2ce1f41 service nova] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.100876] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.576s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.101965] env[62569]: DEBUG nova.compute.manager [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 681.104362] env[62569]: DEBUG oslo_concurrency.lockutils [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.681s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.129392] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Acquiring lock "refresh_cache-0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.370683] env[62569]: DEBUG oslo_concurrency.lockutils [req-ad9acd59-10f7-4789-b147-fc682867d940 req-8b4bf451-3615-467c-baeb-b8d4f2ce1f41 service nova] Releasing lock "refresh_cache-0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.371230] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Acquired lock "refresh_cache-0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.371421] env[62569]: DEBUG nova.network.neutron [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 681.610079] env[62569]: DEBUG nova.compute.utils [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 681.614831] env[62569]: DEBUG nova.compute.manager [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 681.616096] env[62569]: DEBUG nova.network.neutron [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 681.673688] env[62569]: DEBUG nova.policy [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ab6d20862c54432cbafdda33ccc974ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '530ca7e7924743ab91a362a064a3111b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 681.903647] env[62569]: DEBUG nova.network.neutron [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.043616] env[62569]: DEBUG nova.network.neutron [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.046711] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73970227-60b1-4734-8b90-04026e58d19e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.055594] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-459ac675-0269-42e1-b4f0-ee123bf33d5b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.088916] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39492a84-5f87-4c87-aeb9-22f9203bed99 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.099353] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00bdbdda-8eff-48ef-98f4-f2cd1cb8fe79 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.112602] env[62569]: DEBUG nova.compute.provider_tree [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 682.116108] env[62569]: DEBUG nova.compute.manager [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 682.144776] env[62569]: DEBUG nova.network.neutron [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Successfully created port: 691d17e1-efea-40ed-972b-ff3edd53f6af {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 682.248136] env[62569]: DEBUG nova.compute.manager [req-dfba1a74-754e-46a2-801b-32a5986dd608 req-1b6b7622-c84c-4b4e-9995-9bde54bebb08 service nova] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Received event network-vif-deleted-a9cadb21-f24f-480a-819a-10cfdb8e514b {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 682.551679] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Releasing lock "refresh_cache-0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.553250] env[62569]: DEBUG nova.compute.manager [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 682.553475] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 682.553780] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6d8a36c0-5aba-4942-976c-400ed69d6039 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.566463] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24af4fd2-d2bc-4177-a945-db7ee723c5f8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.588125] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97 could not be found. [ 682.588354] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 682.588529] env[62569]: INFO nova.compute.manager [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Took 0.04 seconds to destroy the instance on the hypervisor. [ 682.588771] env[62569]: DEBUG oslo.service.loopingcall [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 682.588989] env[62569]: DEBUG nova.compute.manager [-] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 682.589534] env[62569]: DEBUG nova.network.neutron [-] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 682.613953] env[62569]: DEBUG nova.network.neutron [-] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.621475] env[62569]: DEBUG nova.scheduler.client.report [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 683.116972] env[62569]: DEBUG nova.network.neutron [-] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.129255] env[62569]: DEBUG oslo_concurrency.lockutils [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.025s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.129898] env[62569]: ERROR nova.compute.manager [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 30a8edfc-9961-47b8-910a-626126d325e7, please check neutron logs for more information. [ 683.129898] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Traceback (most recent call last): [ 683.129898] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 683.129898] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] self.driver.spawn(context, instance, image_meta, [ 683.129898] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 683.129898] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 683.129898] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 683.129898] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] vm_ref = self.build_virtual_machine(instance, [ 683.129898] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 683.129898] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] vif_infos = vmwarevif.get_vif_info(self._session, [ 683.129898] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 683.130240] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] for vif in network_info: [ 683.130240] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 683.130240] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] return self._sync_wrapper(fn, *args, **kwargs) [ 683.130240] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 683.130240] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] self.wait() [ 683.130240] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 683.130240] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] self[:] = self._gt.wait() [ 683.130240] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 683.130240] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] return self._exit_event.wait() [ 683.130240] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 683.130240] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] result = hub.switch() [ 683.130240] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 683.130240] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] return self.greenlet.switch() [ 683.130578] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 683.130578] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] result = function(*args, **kwargs) [ 683.130578] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 683.130578] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] return func(*args, **kwargs) [ 683.130578] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 683.130578] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] raise e [ 683.130578] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 683.130578] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] nwinfo = self.network_api.allocate_for_instance( [ 683.130578] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 683.130578] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] created_port_ids = self._update_ports_for_instance( [ 683.130578] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 683.130578] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] with excutils.save_and_reraise_exception(): [ 683.130578] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.130922] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] self.force_reraise() [ 683.130922] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.130922] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] raise self.value [ 683.130922] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 683.130922] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] updated_port = self._update_port( [ 683.130922] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.130922] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] _ensure_no_port_binding_failure(port) [ 683.130922] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.130922] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] raise exception.PortBindingFailed(port_id=port['id']) [ 683.130922] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] nova.exception.PortBindingFailed: Binding failed for port 30a8edfc-9961-47b8-910a-626126d325e7, please check neutron logs for more information. [ 683.130922] env[62569]: ERROR nova.compute.manager [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] [ 683.131303] env[62569]: DEBUG nova.compute.utils [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Binding failed for port 30a8edfc-9961-47b8-910a-626126d325e7, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 683.132173] env[62569]: DEBUG nova.compute.manager [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Build of instance 7cb248f2-2db2-4f62-b2d1-332488170c9b was re-scheduled: Binding failed for port 30a8edfc-9961-47b8-910a-626126d325e7, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 683.132593] env[62569]: DEBUG nova.compute.manager [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 683.132820] env[62569]: DEBUG oslo_concurrency.lockutils [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Acquiring lock "refresh_cache-7cb248f2-2db2-4f62-b2d1-332488170c9b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.132986] env[62569]: DEBUG oslo_concurrency.lockutils [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Acquired lock "refresh_cache-7cb248f2-2db2-4f62-b2d1-332488170c9b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.133165] env[62569]: DEBUG nova.network.neutron [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 683.134047] env[62569]: DEBUG oslo_concurrency.lockutils [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.202s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.135630] env[62569]: INFO nova.compute.claims [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 683.139312] env[62569]: DEBUG nova.compute.manager [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 683.162099] env[62569]: DEBUG nova.network.neutron [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.168714] env[62569]: DEBUG nova.virt.hardware [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 683.168952] env[62569]: DEBUG nova.virt.hardware [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 683.169122] env[62569]: DEBUG nova.virt.hardware [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 683.169305] env[62569]: DEBUG nova.virt.hardware [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 683.169550] env[62569]: DEBUG nova.virt.hardware [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 683.169750] env[62569]: DEBUG nova.virt.hardware [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 683.169837] env[62569]: DEBUG nova.virt.hardware [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 683.169989] env[62569]: DEBUG nova.virt.hardware [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 683.172420] env[62569]: DEBUG nova.virt.hardware [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 683.172420] env[62569]: DEBUG nova.virt.hardware [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 683.172420] env[62569]: DEBUG nova.virt.hardware [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 683.174164] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c48e015-7044-46cd-845f-f209f7b95758 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.183690] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dff5c3d-a42c-4ee1-8482-ac246aa73b31 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.204282] env[62569]: ERROR nova.compute.manager [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 691d17e1-efea-40ed-972b-ff3edd53f6af, please check neutron logs for more information. [ 683.204282] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 683.204282] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 683.204282] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 683.204282] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 683.204282] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 683.204282] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 683.204282] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 683.204282] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.204282] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 683.204282] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.204282] env[62569]: ERROR nova.compute.manager raise self.value [ 683.204282] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 683.204282] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 683.204282] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.204282] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 683.204786] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.204786] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 683.204786] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 691d17e1-efea-40ed-972b-ff3edd53f6af, please check neutron logs for more information. [ 683.204786] env[62569]: ERROR nova.compute.manager [ 683.204786] env[62569]: Traceback (most recent call last): [ 683.204786] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 683.204786] env[62569]: listener.cb(fileno) [ 683.204786] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 683.204786] env[62569]: result = function(*args, **kwargs) [ 683.204786] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 683.204786] env[62569]: return func(*args, **kwargs) [ 683.204786] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 683.204786] env[62569]: raise e [ 683.204786] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 683.204786] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 683.204786] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 683.204786] env[62569]: created_port_ids = self._update_ports_for_instance( [ 683.204786] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 683.204786] env[62569]: with excutils.save_and_reraise_exception(): [ 683.204786] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.204786] env[62569]: self.force_reraise() [ 683.204786] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.204786] env[62569]: raise self.value [ 683.204786] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 683.204786] env[62569]: updated_port = self._update_port( [ 683.204786] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.204786] env[62569]: _ensure_no_port_binding_failure(port) [ 683.204786] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.204786] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 683.205711] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 691d17e1-efea-40ed-972b-ff3edd53f6af, please check neutron logs for more information. [ 683.205711] env[62569]: Removing descriptor: 17 [ 683.205711] env[62569]: ERROR nova.compute.manager [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 691d17e1-efea-40ed-972b-ff3edd53f6af, please check neutron logs for more information. [ 683.205711] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Traceback (most recent call last): [ 683.205711] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 683.205711] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] yield resources [ 683.205711] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 683.205711] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] self.driver.spawn(context, instance, image_meta, [ 683.205711] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 683.205711] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 683.205711] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 683.205711] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] vm_ref = self.build_virtual_machine(instance, [ 683.206080] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 683.206080] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] vif_infos = vmwarevif.get_vif_info(self._session, [ 683.206080] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 683.206080] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] for vif in network_info: [ 683.206080] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 683.206080] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] return self._sync_wrapper(fn, *args, **kwargs) [ 683.206080] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 683.206080] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] self.wait() [ 683.206080] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 683.206080] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] self[:] = self._gt.wait() [ 683.206080] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 683.206080] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] return self._exit_event.wait() [ 683.206080] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 683.206472] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] result = hub.switch() [ 683.206472] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 683.206472] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] return self.greenlet.switch() [ 683.206472] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 683.206472] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] result = function(*args, **kwargs) [ 683.206472] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 683.206472] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] return func(*args, **kwargs) [ 683.206472] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 683.206472] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] raise e [ 683.206472] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 683.206472] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] nwinfo = self.network_api.allocate_for_instance( [ 683.206472] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 683.206472] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] created_port_ids = self._update_ports_for_instance( [ 683.206957] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 683.206957] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] with excutils.save_and_reraise_exception(): [ 683.206957] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.206957] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] self.force_reraise() [ 683.206957] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.206957] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] raise self.value [ 683.206957] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 683.206957] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] updated_port = self._update_port( [ 683.206957] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.206957] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] _ensure_no_port_binding_failure(port) [ 683.206957] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.206957] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] raise exception.PortBindingFailed(port_id=port['id']) [ 683.207336] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] nova.exception.PortBindingFailed: Binding failed for port 691d17e1-efea-40ed-972b-ff3edd53f6af, please check neutron logs for more information. [ 683.207336] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] [ 683.207336] env[62569]: INFO nova.compute.manager [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Terminating instance [ 683.233549] env[62569]: DEBUG nova.network.neutron [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.620663] env[62569]: INFO nova.compute.manager [-] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Took 1.03 seconds to deallocate network for instance. [ 683.623068] env[62569]: DEBUG nova.compute.claims [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 683.623253] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.710743] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "refresh_cache-c3f25b97-0e1e-4e78-94f0-e8087d1b55c1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.710973] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "refresh_cache-c3f25b97-0e1e-4e78-94f0-e8087d1b55c1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.711176] env[62569]: DEBUG nova.network.neutron [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 683.735530] env[62569]: DEBUG oslo_concurrency.lockutils [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Releasing lock "refresh_cache-7cb248f2-2db2-4f62-b2d1-332488170c9b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.735795] env[62569]: DEBUG nova.compute.manager [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 683.736011] env[62569]: DEBUG nova.compute.manager [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 683.736162] env[62569]: DEBUG nova.network.neutron [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 683.752506] env[62569]: DEBUG nova.network.neutron [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.229009] env[62569]: DEBUG nova.network.neutron [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.254947] env[62569]: DEBUG nova.network.neutron [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.277082] env[62569]: DEBUG nova.compute.manager [req-4cbedbbc-adb3-4b79-8a1e-ad330c9d5070 req-9db863a4-a1ce-4bf8-99dd-d57ac02b8163 service nova] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Received event network-changed-691d17e1-efea-40ed-972b-ff3edd53f6af {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 684.277277] env[62569]: DEBUG nova.compute.manager [req-4cbedbbc-adb3-4b79-8a1e-ad330c9d5070 req-9db863a4-a1ce-4bf8-99dd-d57ac02b8163 service nova] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Refreshing instance network info cache due to event network-changed-691d17e1-efea-40ed-972b-ff3edd53f6af. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 684.277468] env[62569]: DEBUG oslo_concurrency.lockutils [req-4cbedbbc-adb3-4b79-8a1e-ad330c9d5070 req-9db863a4-a1ce-4bf8-99dd-d57ac02b8163 service nova] Acquiring lock "refresh_cache-c3f25b97-0e1e-4e78-94f0-e8087d1b55c1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.318858] env[62569]: DEBUG nova.network.neutron [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.498709] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b667f9f-3a1d-4968-8dc3-99fcd780ae9d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.506736] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4511f082-aa32-4728-8393-d0dabfc20c67 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.537283] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85dc6041-7262-4021-910f-67b427504b7e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.544951] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52b3abd0-3691-4100-b79d-e5dce9c52559 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.558905] env[62569]: DEBUG nova.compute.provider_tree [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 684.758409] env[62569]: INFO nova.compute.manager [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] [instance: 7cb248f2-2db2-4f62-b2d1-332488170c9b] Took 1.02 seconds to deallocate network for instance. [ 684.822090] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "refresh_cache-c3f25b97-0e1e-4e78-94f0-e8087d1b55c1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.822547] env[62569]: DEBUG nova.compute.manager [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 684.822742] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 684.823328] env[62569]: DEBUG oslo_concurrency.lockutils [req-4cbedbbc-adb3-4b79-8a1e-ad330c9d5070 req-9db863a4-a1ce-4bf8-99dd-d57ac02b8163 service nova] Acquired lock "refresh_cache-c3f25b97-0e1e-4e78-94f0-e8087d1b55c1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.823328] env[62569]: DEBUG nova.network.neutron [req-4cbedbbc-adb3-4b79-8a1e-ad330c9d5070 req-9db863a4-a1ce-4bf8-99dd-d57ac02b8163 service nova] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Refreshing network info cache for port 691d17e1-efea-40ed-972b-ff3edd53f6af {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 684.824209] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c941d03b-f2b4-48cb-96d3-6309243f7542 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.834196] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d52fe6e-af44-4196-b786-97a23fb18cac {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.855053] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c3f25b97-0e1e-4e78-94f0-e8087d1b55c1 could not be found. [ 684.855322] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 684.855506] env[62569]: INFO nova.compute.manager [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Took 0.03 seconds to destroy the instance on the hypervisor. [ 684.855746] env[62569]: DEBUG oslo.service.loopingcall [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 684.855972] env[62569]: DEBUG nova.compute.manager [-] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 684.856091] env[62569]: DEBUG nova.network.neutron [-] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 684.870106] env[62569]: DEBUG nova.network.neutron [-] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.062677] env[62569]: DEBUG nova.scheduler.client.report [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 685.341798] env[62569]: DEBUG nova.network.neutron [req-4cbedbbc-adb3-4b79-8a1e-ad330c9d5070 req-9db863a4-a1ce-4bf8-99dd-d57ac02b8163 service nova] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.372528] env[62569]: DEBUG nova.network.neutron [-] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.411804] env[62569]: DEBUG nova.network.neutron [req-4cbedbbc-adb3-4b79-8a1e-ad330c9d5070 req-9db863a4-a1ce-4bf8-99dd-d57ac02b8163 service nova] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.568118] env[62569]: DEBUG oslo_concurrency.lockutils [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.434s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.568640] env[62569]: DEBUG nova.compute.manager [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 685.571631] env[62569]: DEBUG oslo_concurrency.lockutils [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.882s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.787181] env[62569]: INFO nova.scheduler.client.report [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Deleted allocations for instance 7cb248f2-2db2-4f62-b2d1-332488170c9b [ 685.874881] env[62569]: INFO nova.compute.manager [-] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Took 1.02 seconds to deallocate network for instance. [ 685.879848] env[62569]: DEBUG nova.compute.claims [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 685.880068] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.914944] env[62569]: DEBUG oslo_concurrency.lockutils [req-4cbedbbc-adb3-4b79-8a1e-ad330c9d5070 req-9db863a4-a1ce-4bf8-99dd-d57ac02b8163 service nova] Releasing lock "refresh_cache-c3f25b97-0e1e-4e78-94f0-e8087d1b55c1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.915194] env[62569]: DEBUG nova.compute.manager [req-4cbedbbc-adb3-4b79-8a1e-ad330c9d5070 req-9db863a4-a1ce-4bf8-99dd-d57ac02b8163 service nova] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Received event network-vif-deleted-691d17e1-efea-40ed-972b-ff3edd53f6af {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 686.076077] env[62569]: DEBUG nova.compute.utils [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 686.080257] env[62569]: DEBUG nova.compute.manager [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 686.080429] env[62569]: DEBUG nova.network.neutron [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 686.129469] env[62569]: DEBUG nova.policy [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '98e5b0ac5257414db011e1bc7ff76468', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '52313891dc7544d7ad788a9cd64b4a03', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 686.295434] env[62569]: DEBUG oslo_concurrency.lockutils [None req-377581d6-3023-4c49-93fa-7f8ec5dda911 tempest-MigrationsAdminTest-1614805254 tempest-MigrationsAdminTest-1614805254-project-member] Lock "7cb248f2-2db2-4f62-b2d1-332488170c9b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 129.772s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.437615] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1648ed22-cfb6-471c-a99a-7c995a058ba4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.442296] env[62569]: DEBUG nova.network.neutron [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Successfully created port: 99b36417-48ee-4881-a950-cfca6217ccb8 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 686.449286] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-648077ef-0901-4d5b-9484-c5abf422c2b1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.483357] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f0ffe99-4e9a-48e1-bf1b-03fcd8ba2b99 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.491126] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34bd963d-21ba-4703-94ca-a29e1be074c4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.504048] env[62569]: DEBUG nova.compute.provider_tree [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.583644] env[62569]: DEBUG nova.compute.manager [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 686.800165] env[62569]: DEBUG nova.compute.manager [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 687.009360] env[62569]: DEBUG nova.scheduler.client.report [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 687.213254] env[62569]: DEBUG nova.compute.manager [req-28b03aa6-7e1d-4f58-9459-b52aa5eca36f req-34fde54e-5b6c-44ff-9ef3-f48a9c9862f0 service nova] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Received event network-changed-99b36417-48ee-4881-a950-cfca6217ccb8 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 687.213254] env[62569]: DEBUG nova.compute.manager [req-28b03aa6-7e1d-4f58-9459-b52aa5eca36f req-34fde54e-5b6c-44ff-9ef3-f48a9c9862f0 service nova] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Refreshing instance network info cache due to event network-changed-99b36417-48ee-4881-a950-cfca6217ccb8. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 687.213254] env[62569]: DEBUG oslo_concurrency.lockutils [req-28b03aa6-7e1d-4f58-9459-b52aa5eca36f req-34fde54e-5b6c-44ff-9ef3-f48a9c9862f0 service nova] Acquiring lock "refresh_cache-d3274d59-6cf7-4e86-90f9-ffea49ad5342" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.213254] env[62569]: DEBUG oslo_concurrency.lockutils [req-28b03aa6-7e1d-4f58-9459-b52aa5eca36f req-34fde54e-5b6c-44ff-9ef3-f48a9c9862f0 service nova] Acquired lock "refresh_cache-d3274d59-6cf7-4e86-90f9-ffea49ad5342" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.213254] env[62569]: DEBUG nova.network.neutron [req-28b03aa6-7e1d-4f58-9459-b52aa5eca36f req-34fde54e-5b6c-44ff-9ef3-f48a9c9862f0 service nova] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Refreshing network info cache for port 99b36417-48ee-4881-a950-cfca6217ccb8 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 687.325237] env[62569]: DEBUG oslo_concurrency.lockutils [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.399847] env[62569]: ERROR nova.compute.manager [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 99b36417-48ee-4881-a950-cfca6217ccb8, please check neutron logs for more information. [ 687.399847] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 687.399847] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 687.399847] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 687.399847] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.399847] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 687.399847] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.399847] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 687.399847] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.399847] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 687.399847] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.399847] env[62569]: ERROR nova.compute.manager raise self.value [ 687.399847] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.399847] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 687.399847] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.399847] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 687.400518] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.400518] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 687.400518] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 99b36417-48ee-4881-a950-cfca6217ccb8, please check neutron logs for more information. [ 687.400518] env[62569]: ERROR nova.compute.manager [ 687.400518] env[62569]: Traceback (most recent call last): [ 687.400518] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 687.400518] env[62569]: listener.cb(fileno) [ 687.400518] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.400518] env[62569]: result = function(*args, **kwargs) [ 687.400518] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 687.400518] env[62569]: return func(*args, **kwargs) [ 687.400518] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 687.400518] env[62569]: raise e [ 687.400518] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 687.400518] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 687.400518] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.400518] env[62569]: created_port_ids = self._update_ports_for_instance( [ 687.400518] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.400518] env[62569]: with excutils.save_and_reraise_exception(): [ 687.400518] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.400518] env[62569]: self.force_reraise() [ 687.400518] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.400518] env[62569]: raise self.value [ 687.400518] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.400518] env[62569]: updated_port = self._update_port( [ 687.400518] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.400518] env[62569]: _ensure_no_port_binding_failure(port) [ 687.400518] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.400518] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 687.401686] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 99b36417-48ee-4881-a950-cfca6217ccb8, please check neutron logs for more information. [ 687.401686] env[62569]: Removing descriptor: 14 [ 687.514783] env[62569]: DEBUG oslo_concurrency.lockutils [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.943s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.515203] env[62569]: ERROR nova.compute.manager [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a0f99f88-8014-4300-863a-38e81bc640d2, please check neutron logs for more information. [ 687.515203] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Traceback (most recent call last): [ 687.515203] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 687.515203] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] self.driver.spawn(context, instance, image_meta, [ 687.515203] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 687.515203] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 687.515203] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 687.515203] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] vm_ref = self.build_virtual_machine(instance, [ 687.515203] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 687.515203] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] vif_infos = vmwarevif.get_vif_info(self._session, [ 687.515203] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 687.515566] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] for vif in network_info: [ 687.515566] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 687.515566] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] return self._sync_wrapper(fn, *args, **kwargs) [ 687.515566] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 687.515566] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] self.wait() [ 687.515566] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 687.515566] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] self[:] = self._gt.wait() [ 687.515566] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 687.515566] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] return self._exit_event.wait() [ 687.515566] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 687.515566] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] current.throw(*self._exc) [ 687.515566] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.515566] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] result = function(*args, **kwargs) [ 687.515879] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 687.515879] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] return func(*args, **kwargs) [ 687.515879] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 687.515879] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] raise e [ 687.515879] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 687.515879] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] nwinfo = self.network_api.allocate_for_instance( [ 687.515879] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.515879] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] created_port_ids = self._update_ports_for_instance( [ 687.515879] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.515879] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] with excutils.save_and_reraise_exception(): [ 687.515879] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.515879] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] self.force_reraise() [ 687.515879] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.516203] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] raise self.value [ 687.516203] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.516203] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] updated_port = self._update_port( [ 687.516203] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.516203] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] _ensure_no_port_binding_failure(port) [ 687.516203] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.516203] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] raise exception.PortBindingFailed(port_id=port['id']) [ 687.516203] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] nova.exception.PortBindingFailed: Binding failed for port a0f99f88-8014-4300-863a-38e81bc640d2, please check neutron logs for more information. [ 687.516203] env[62569]: ERROR nova.compute.manager [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] [ 687.516203] env[62569]: DEBUG nova.compute.utils [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Binding failed for port a0f99f88-8014-4300-863a-38e81bc640d2, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 687.517157] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.161s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.518618] env[62569]: INFO nova.compute.claims [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 687.521184] env[62569]: DEBUG nova.compute.manager [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Build of instance 695022e4-bb58-44d7-87e5-58bbd577ec7b was re-scheduled: Binding failed for port a0f99f88-8014-4300-863a-38e81bc640d2, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 687.521598] env[62569]: DEBUG nova.compute.manager [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 687.521820] env[62569]: DEBUG oslo_concurrency.lockutils [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquiring lock "refresh_cache-695022e4-bb58-44d7-87e5-58bbd577ec7b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.521977] env[62569]: DEBUG oslo_concurrency.lockutils [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquired lock "refresh_cache-695022e4-bb58-44d7-87e5-58bbd577ec7b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.522212] env[62569]: DEBUG nova.network.neutron [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 687.597796] env[62569]: DEBUG nova.compute.manager [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 687.621088] env[62569]: DEBUG nova.virt.hardware [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 687.621548] env[62569]: DEBUG nova.virt.hardware [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 687.621823] env[62569]: DEBUG nova.virt.hardware [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 687.622129] env[62569]: DEBUG nova.virt.hardware [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 687.622477] env[62569]: DEBUG nova.virt.hardware [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 687.622738] env[62569]: DEBUG nova.virt.hardware [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 687.623059] env[62569]: DEBUG nova.virt.hardware [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 687.623328] env[62569]: DEBUG nova.virt.hardware [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 687.623607] env[62569]: DEBUG nova.virt.hardware [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 687.623895] env[62569]: DEBUG nova.virt.hardware [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 687.624189] env[62569]: DEBUG nova.virt.hardware [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 687.625504] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b0ebcdd-1eda-47b6-b74a-f3448dbd8f22 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.633830] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1037d47a-35a4-438e-938e-c9d225c3bbbc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.649134] env[62569]: ERROR nova.compute.manager [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 99b36417-48ee-4881-a950-cfca6217ccb8, please check neutron logs for more information. [ 687.649134] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Traceback (most recent call last): [ 687.649134] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 687.649134] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] yield resources [ 687.649134] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 687.649134] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] self.driver.spawn(context, instance, image_meta, [ 687.649134] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 687.649134] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] self._vmops.spawn(context, instance, image_meta, injected_files, [ 687.649134] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 687.649134] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] vm_ref = self.build_virtual_machine(instance, [ 687.649134] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 687.649470] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] vif_infos = vmwarevif.get_vif_info(self._session, [ 687.649470] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 687.649470] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] for vif in network_info: [ 687.649470] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 687.649470] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] return self._sync_wrapper(fn, *args, **kwargs) [ 687.649470] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 687.649470] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] self.wait() [ 687.649470] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 687.649470] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] self[:] = self._gt.wait() [ 687.649470] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 687.649470] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] return self._exit_event.wait() [ 687.649470] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 687.649470] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] current.throw(*self._exc) [ 687.649801] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.649801] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] result = function(*args, **kwargs) [ 687.649801] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 687.649801] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] return func(*args, **kwargs) [ 687.649801] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 687.649801] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] raise e [ 687.649801] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 687.649801] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] nwinfo = self.network_api.allocate_for_instance( [ 687.649801] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.649801] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] created_port_ids = self._update_ports_for_instance( [ 687.649801] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.649801] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] with excutils.save_and_reraise_exception(): [ 687.649801] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.650164] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] self.force_reraise() [ 687.650164] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.650164] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] raise self.value [ 687.650164] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.650164] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] updated_port = self._update_port( [ 687.650164] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.650164] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] _ensure_no_port_binding_failure(port) [ 687.650164] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.650164] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] raise exception.PortBindingFailed(port_id=port['id']) [ 687.650164] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] nova.exception.PortBindingFailed: Binding failed for port 99b36417-48ee-4881-a950-cfca6217ccb8, please check neutron logs for more information. [ 687.650164] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] [ 687.652026] env[62569]: INFO nova.compute.manager [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Terminating instance [ 687.735849] env[62569]: DEBUG nova.network.neutron [req-28b03aa6-7e1d-4f58-9459-b52aa5eca36f req-34fde54e-5b6c-44ff-9ef3-f48a9c9862f0 service nova] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 687.820854] env[62569]: DEBUG nova.network.neutron [req-28b03aa6-7e1d-4f58-9459-b52aa5eca36f req-34fde54e-5b6c-44ff-9ef3-f48a9c9862f0 service nova] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.045130] env[62569]: DEBUG nova.network.neutron [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.123218] env[62569]: DEBUG nova.network.neutron [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.154614] env[62569]: DEBUG oslo_concurrency.lockutils [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Acquiring lock "refresh_cache-d3274d59-6cf7-4e86-90f9-ffea49ad5342" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.323780] env[62569]: DEBUG oslo_concurrency.lockutils [req-28b03aa6-7e1d-4f58-9459-b52aa5eca36f req-34fde54e-5b6c-44ff-9ef3-f48a9c9862f0 service nova] Releasing lock "refresh_cache-d3274d59-6cf7-4e86-90f9-ffea49ad5342" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.323780] env[62569]: DEBUG oslo_concurrency.lockutils [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Acquired lock "refresh_cache-d3274d59-6cf7-4e86-90f9-ffea49ad5342" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.324087] env[62569]: DEBUG nova.network.neutron [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 688.626239] env[62569]: DEBUG oslo_concurrency.lockutils [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Releasing lock "refresh_cache-695022e4-bb58-44d7-87e5-58bbd577ec7b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.626474] env[62569]: DEBUG nova.compute.manager [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 688.626654] env[62569]: DEBUG nova.compute.manager [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 688.626823] env[62569]: DEBUG nova.network.neutron [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 688.643118] env[62569]: DEBUG nova.network.neutron [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.858059] env[62569]: DEBUG nova.network.neutron [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.883233] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa5d4d98-4709-477e-a505-6555aaea9f29 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.892018] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9a33ea9-21ed-422f-890b-da214a84486b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.926834] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dc5cd35-cd92-4d01-a645-c2f144298e2e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.935973] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e82bc47c-3601-4c36-be7f-674af101dcc3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.950750] env[62569]: DEBUG nova.compute.provider_tree [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 688.994704] env[62569]: DEBUG nova.network.neutron [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.148919] env[62569]: DEBUG nova.network.neutron [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.272496] env[62569]: DEBUG nova.compute.manager [req-6ce9f6f6-77c9-4289-a248-3079e8fbd45d req-5051d852-2918-4554-93eb-2d8c327b5c42 service nova] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Received event network-vif-deleted-99b36417-48ee-4881-a950-cfca6217ccb8 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 689.452516] env[62569]: DEBUG nova.scheduler.client.report [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 689.502017] env[62569]: DEBUG oslo_concurrency.lockutils [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Releasing lock "refresh_cache-d3274d59-6cf7-4e86-90f9-ffea49ad5342" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.502452] env[62569]: DEBUG nova.compute.manager [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 689.502662] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 689.502970] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c5bac8e7-affb-41d5-ac11-87120b21b927 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.514028] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a12ec0ec-3a56-4bb4-8e04-2e39e8840536 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.539263] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d3274d59-6cf7-4e86-90f9-ffea49ad5342 could not be found. [ 689.539519] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 689.539713] env[62569]: INFO nova.compute.manager [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Took 0.04 seconds to destroy the instance on the hypervisor. [ 689.539984] env[62569]: DEBUG oslo.service.loopingcall [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 689.540260] env[62569]: DEBUG nova.compute.manager [-] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 689.540379] env[62569]: DEBUG nova.network.neutron [-] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 689.560222] env[62569]: DEBUG nova.network.neutron [-] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.651380] env[62569]: INFO nova.compute.manager [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 695022e4-bb58-44d7-87e5-58bbd577ec7b] Took 1.02 seconds to deallocate network for instance. [ 689.962019] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.444s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.962019] env[62569]: DEBUG nova.compute.manager [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 689.964580] env[62569]: DEBUG oslo_concurrency.lockutils [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.561s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.965951] env[62569]: INFO nova.compute.claims [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 690.062648] env[62569]: DEBUG nova.network.neutron [-] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.470915] env[62569]: DEBUG nova.compute.utils [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 690.474039] env[62569]: DEBUG nova.compute.manager [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 690.474214] env[62569]: DEBUG nova.network.neutron [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 690.518558] env[62569]: DEBUG nova.policy [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0a9d5c0ca83b488e9bcaec31c7867b44', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '17a68974796a49359e6498789537867a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 690.565032] env[62569]: INFO nova.compute.manager [-] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Took 1.02 seconds to deallocate network for instance. [ 690.569753] env[62569]: DEBUG nova.compute.claims [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 690.570597] env[62569]: DEBUG oslo_concurrency.lockutils [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.690305] env[62569]: INFO nova.scheduler.client.report [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Deleted allocations for instance 695022e4-bb58-44d7-87e5-58bbd577ec7b [ 690.765219] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Acquiring lock "b3ccc28b-68d0-461d-a67d-b5ad179a80f9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.765219] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Lock "b3ccc28b-68d0-461d-a67d-b5ad179a80f9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.810090] env[62569]: DEBUG nova.network.neutron [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Successfully created port: decaf5e6-d463-4604-a2f9-3c72e5f71a6e {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 690.976021] env[62569]: DEBUG nova.compute.manager [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 691.200749] env[62569]: DEBUG oslo_concurrency.lockutils [None req-943a5398-d7dd-45dd-959d-938b9f581e59 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "695022e4-bb58-44d7-87e5-58bbd577ec7b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.106s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.371459] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-790f446d-e455-411e-bab9-b7a2c5fc7706 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.381043] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43066bfe-7b10-45d5-a5eb-cf937d3ef6f8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.414954] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d687818-9bb7-440c-a8fd-f7642d722396 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.423207] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-422e0dd0-7159-4a0a-9c81-de5076ee4d31 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.436507] env[62569]: DEBUG nova.compute.provider_tree [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 691.609868] env[62569]: DEBUG nova.compute.manager [req-91451cab-4adf-4c67-bc0b-259e9b63cdf2 req-c89326cb-cc2c-451c-8984-e3be96ba5d9c service nova] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Received event network-changed-decaf5e6-d463-4604-a2f9-3c72e5f71a6e {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 691.610162] env[62569]: DEBUG nova.compute.manager [req-91451cab-4adf-4c67-bc0b-259e9b63cdf2 req-c89326cb-cc2c-451c-8984-e3be96ba5d9c service nova] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Refreshing instance network info cache due to event network-changed-decaf5e6-d463-4604-a2f9-3c72e5f71a6e. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 691.610484] env[62569]: DEBUG oslo_concurrency.lockutils [req-91451cab-4adf-4c67-bc0b-259e9b63cdf2 req-c89326cb-cc2c-451c-8984-e3be96ba5d9c service nova] Acquiring lock "refresh_cache-4f45ccc2-0352-4c25-9a92-aee8f051c0b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.610675] env[62569]: DEBUG oslo_concurrency.lockutils [req-91451cab-4adf-4c67-bc0b-259e9b63cdf2 req-c89326cb-cc2c-451c-8984-e3be96ba5d9c service nova] Acquired lock "refresh_cache-4f45ccc2-0352-4c25-9a92-aee8f051c0b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.610831] env[62569]: DEBUG nova.network.neutron [req-91451cab-4adf-4c67-bc0b-259e9b63cdf2 req-c89326cb-cc2c-451c-8984-e3be96ba5d9c service nova] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Refreshing network info cache for port decaf5e6-d463-4604-a2f9-3c72e5f71a6e {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 691.694315] env[62569]: ERROR nova.compute.manager [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port decaf5e6-d463-4604-a2f9-3c72e5f71a6e, please check neutron logs for more information. [ 691.694315] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 691.694315] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 691.694315] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 691.694315] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 691.694315] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 691.694315] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 691.694315] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 691.694315] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.694315] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 691.694315] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.694315] env[62569]: ERROR nova.compute.manager raise self.value [ 691.694315] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 691.694315] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 691.694315] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.694315] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 691.694925] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.694925] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 691.694925] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port decaf5e6-d463-4604-a2f9-3c72e5f71a6e, please check neutron logs for more information. [ 691.694925] env[62569]: ERROR nova.compute.manager [ 691.694925] env[62569]: Traceback (most recent call last): [ 691.694925] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 691.694925] env[62569]: listener.cb(fileno) [ 691.694925] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 691.694925] env[62569]: result = function(*args, **kwargs) [ 691.694925] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 691.694925] env[62569]: return func(*args, **kwargs) [ 691.694925] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 691.694925] env[62569]: raise e [ 691.694925] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 691.694925] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 691.694925] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 691.694925] env[62569]: created_port_ids = self._update_ports_for_instance( [ 691.694925] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 691.694925] env[62569]: with excutils.save_and_reraise_exception(): [ 691.694925] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.694925] env[62569]: self.force_reraise() [ 691.694925] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.694925] env[62569]: raise self.value [ 691.694925] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 691.694925] env[62569]: updated_port = self._update_port( [ 691.694925] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.694925] env[62569]: _ensure_no_port_binding_failure(port) [ 691.694925] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.694925] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 691.695899] env[62569]: nova.exception.PortBindingFailed: Binding failed for port decaf5e6-d463-4604-a2f9-3c72e5f71a6e, please check neutron logs for more information. [ 691.695899] env[62569]: Removing descriptor: 14 [ 691.706206] env[62569]: DEBUG nova.compute.manager [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 691.940615] env[62569]: DEBUG nova.scheduler.client.report [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 691.995571] env[62569]: DEBUG nova.compute.manager [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 692.029197] env[62569]: DEBUG nova.virt.hardware [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 692.029433] env[62569]: DEBUG nova.virt.hardware [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 692.029632] env[62569]: DEBUG nova.virt.hardware [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 692.029866] env[62569]: DEBUG nova.virt.hardware [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 692.029991] env[62569]: DEBUG nova.virt.hardware [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 692.030117] env[62569]: DEBUG nova.virt.hardware [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 692.030327] env[62569]: DEBUG nova.virt.hardware [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 692.031766] env[62569]: DEBUG nova.virt.hardware [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 692.031766] env[62569]: DEBUG nova.virt.hardware [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 692.031766] env[62569]: DEBUG nova.virt.hardware [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 692.031766] env[62569]: DEBUG nova.virt.hardware [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 692.032182] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d347e37d-fd8a-4e6a-8a9c-e8e1081a3fc0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.042494] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e928eb1e-9ad8-432c-87c8-4b2f87a8b789 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.057731] env[62569]: ERROR nova.compute.manager [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port decaf5e6-d463-4604-a2f9-3c72e5f71a6e, please check neutron logs for more information. [ 692.057731] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Traceback (most recent call last): [ 692.057731] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 692.057731] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] yield resources [ 692.057731] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 692.057731] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] self.driver.spawn(context, instance, image_meta, [ 692.057731] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 692.057731] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 692.057731] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 692.057731] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] vm_ref = self.build_virtual_machine(instance, [ 692.057731] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 692.058157] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] vif_infos = vmwarevif.get_vif_info(self._session, [ 692.058157] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 692.058157] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] for vif in network_info: [ 692.058157] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 692.058157] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] return self._sync_wrapper(fn, *args, **kwargs) [ 692.058157] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 692.058157] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] self.wait() [ 692.058157] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 692.058157] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] self[:] = self._gt.wait() [ 692.058157] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 692.058157] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] return self._exit_event.wait() [ 692.058157] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 692.058157] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] current.throw(*self._exc) [ 692.058550] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 692.058550] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] result = function(*args, **kwargs) [ 692.058550] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 692.058550] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] return func(*args, **kwargs) [ 692.058550] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 692.058550] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] raise e [ 692.058550] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 692.058550] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] nwinfo = self.network_api.allocate_for_instance( [ 692.058550] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 692.058550] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] created_port_ids = self._update_ports_for_instance( [ 692.058550] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 692.058550] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] with excutils.save_and_reraise_exception(): [ 692.058550] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.058932] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] self.force_reraise() [ 692.058932] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.058932] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] raise self.value [ 692.058932] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 692.058932] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] updated_port = self._update_port( [ 692.058932] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.058932] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] _ensure_no_port_binding_failure(port) [ 692.058932] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.058932] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] raise exception.PortBindingFailed(port_id=port['id']) [ 692.058932] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] nova.exception.PortBindingFailed: Binding failed for port decaf5e6-d463-4604-a2f9-3c72e5f71a6e, please check neutron logs for more information. [ 692.058932] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] [ 692.058932] env[62569]: INFO nova.compute.manager [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Terminating instance [ 692.131704] env[62569]: DEBUG nova.network.neutron [req-91451cab-4adf-4c67-bc0b-259e9b63cdf2 req-c89326cb-cc2c-451c-8984-e3be96ba5d9c service nova] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 692.210767] env[62569]: DEBUG nova.network.neutron [req-91451cab-4adf-4c67-bc0b-259e9b63cdf2 req-c89326cb-cc2c-451c-8984-e3be96ba5d9c service nova] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.235799] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.446530] env[62569]: DEBUG oslo_concurrency.lockutils [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.482s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.447074] env[62569]: DEBUG nova.compute.manager [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 692.451293] env[62569]: DEBUG oslo_concurrency.lockutils [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.404s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.563918] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Acquiring lock "refresh_cache-4f45ccc2-0352-4c25-9a92-aee8f051c0b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.715490] env[62569]: DEBUG oslo_concurrency.lockutils [req-91451cab-4adf-4c67-bc0b-259e9b63cdf2 req-c89326cb-cc2c-451c-8984-e3be96ba5d9c service nova] Releasing lock "refresh_cache-4f45ccc2-0352-4c25-9a92-aee8f051c0b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 692.715922] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Acquired lock "refresh_cache-4f45ccc2-0352-4c25-9a92-aee8f051c0b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.716131] env[62569]: DEBUG nova.network.neutron [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 692.955869] env[62569]: DEBUG nova.compute.utils [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 692.960917] env[62569]: DEBUG nova.compute.manager [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 692.961395] env[62569]: DEBUG nova.network.neutron [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 692.999456] env[62569]: DEBUG nova.policy [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0a9d5c0ca83b488e9bcaec31c7867b44', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '17a68974796a49359e6498789537867a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 693.239566] env[62569]: DEBUG nova.network.neutron [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 693.299044] env[62569]: DEBUG nova.network.neutron [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Successfully created port: fe670c41-ccb2-44ed-bf68-c487c6297401 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 693.317657] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c1e9ec-ddf4-4612-adce-0be84a6e623e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.336845] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93ad74d5-c35b-47d0-9d49-abcb81d20243 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.374797] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-795ace38-b60d-4f31-ba19-cd34a95c0877 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.381398] env[62569]: DEBUG nova.network.neutron [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.388679] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2309e94-35db-4888-9b80-004342ef36a8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.403371] env[62569]: DEBUG nova.compute.provider_tree [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 693.463136] env[62569]: DEBUG nova.compute.manager [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 693.636761] env[62569]: DEBUG nova.compute.manager [req-32377ef1-16f6-485f-9983-868c54221993 req-87e23f42-3be1-40c8-a7ad-2f1906a49825 service nova] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Received event network-vif-deleted-decaf5e6-d463-4604-a2f9-3c72e5f71a6e {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 693.878213] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquiring lock "337d88a2-30b9-4846-929e-042bd7a64a65" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.878556] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "337d88a2-30b9-4846-929e-042bd7a64a65" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.885711] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Releasing lock "refresh_cache-4f45ccc2-0352-4c25-9a92-aee8f051c0b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.886314] env[62569]: DEBUG nova.compute.manager [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 693.886314] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 693.886783] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-47d4fbbc-1edb-4086-92c3-67edb6f85b46 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.895943] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65be8bc0-5d2a-4f19-8cfd-c12962f9b4c1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.907377] env[62569]: DEBUG nova.scheduler.client.report [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 693.927047] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4f45ccc2-0352-4c25-9a92-aee8f051c0b5 could not be found. [ 693.927047] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 693.927047] env[62569]: INFO nova.compute.manager [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 693.927047] env[62569]: DEBUG oslo.service.loopingcall [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 693.929287] env[62569]: DEBUG nova.compute.manager [-] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 693.929287] env[62569]: DEBUG nova.network.neutron [-] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 693.946698] env[62569]: DEBUG nova.network.neutron [-] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.193699] env[62569]: ERROR nova.compute.manager [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fe670c41-ccb2-44ed-bf68-c487c6297401, please check neutron logs for more information. [ 694.193699] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 694.193699] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 694.193699] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 694.193699] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.193699] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 694.193699] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.193699] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 694.193699] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.193699] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 694.193699] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.193699] env[62569]: ERROR nova.compute.manager raise self.value [ 694.193699] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.193699] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 694.193699] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.193699] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 694.194609] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.194609] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 694.194609] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fe670c41-ccb2-44ed-bf68-c487c6297401, please check neutron logs for more information. [ 694.194609] env[62569]: ERROR nova.compute.manager [ 694.194609] env[62569]: Traceback (most recent call last): [ 694.194609] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 694.194609] env[62569]: listener.cb(fileno) [ 694.194609] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.194609] env[62569]: result = function(*args, **kwargs) [ 694.194609] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.194609] env[62569]: return func(*args, **kwargs) [ 694.194609] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 694.194609] env[62569]: raise e [ 694.194609] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 694.194609] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 694.194609] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.194609] env[62569]: created_port_ids = self._update_ports_for_instance( [ 694.194609] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.194609] env[62569]: with excutils.save_and_reraise_exception(): [ 694.194609] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.194609] env[62569]: self.force_reraise() [ 694.194609] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.194609] env[62569]: raise self.value [ 694.194609] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.194609] env[62569]: updated_port = self._update_port( [ 694.194609] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.194609] env[62569]: _ensure_no_port_binding_failure(port) [ 694.194609] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.194609] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 694.195702] env[62569]: nova.exception.PortBindingFailed: Binding failed for port fe670c41-ccb2-44ed-bf68-c487c6297401, please check neutron logs for more information. [ 694.195702] env[62569]: Removing descriptor: 14 [ 694.415474] env[62569]: DEBUG oslo_concurrency.lockutils [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.964s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.416106] env[62569]: ERROR nova.compute.manager [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e4af3f09-3554-4008-bd79-56dd4a9affe7, please check neutron logs for more information. [ 694.416106] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Traceback (most recent call last): [ 694.416106] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 694.416106] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] self.driver.spawn(context, instance, image_meta, [ 694.416106] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 694.416106] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 694.416106] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 694.416106] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] vm_ref = self.build_virtual_machine(instance, [ 694.416106] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 694.416106] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] vif_infos = vmwarevif.get_vif_info(self._session, [ 694.416106] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 694.416468] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] for vif in network_info: [ 694.416468] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 694.416468] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] return self._sync_wrapper(fn, *args, **kwargs) [ 694.416468] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 694.416468] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] self.wait() [ 694.416468] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 694.416468] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] self[:] = self._gt.wait() [ 694.416468] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 694.416468] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] return self._exit_event.wait() [ 694.416468] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 694.416468] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] current.throw(*self._exc) [ 694.416468] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.416468] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] result = function(*args, **kwargs) [ 694.416848] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.416848] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] return func(*args, **kwargs) [ 694.416848] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 694.416848] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] raise e [ 694.416848] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 694.416848] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] nwinfo = self.network_api.allocate_for_instance( [ 694.416848] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.416848] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] created_port_ids = self._update_ports_for_instance( [ 694.416848] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.416848] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] with excutils.save_and_reraise_exception(): [ 694.416848] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.416848] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] self.force_reraise() [ 694.416848] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.417248] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] raise self.value [ 694.417248] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.417248] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] updated_port = self._update_port( [ 694.417248] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.417248] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] _ensure_no_port_binding_failure(port) [ 694.417248] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.417248] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] raise exception.PortBindingFailed(port_id=port['id']) [ 694.417248] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] nova.exception.PortBindingFailed: Binding failed for port e4af3f09-3554-4008-bd79-56dd4a9affe7, please check neutron logs for more information. [ 694.417248] env[62569]: ERROR nova.compute.manager [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] [ 694.417248] env[62569]: DEBUG nova.compute.utils [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Binding failed for port e4af3f09-3554-4008-bd79-56dd4a9affe7, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 694.418559] env[62569]: DEBUG oslo_concurrency.lockutils [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.563s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.420059] env[62569]: INFO nova.compute.claims [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 694.423657] env[62569]: DEBUG nova.compute.manager [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Build of instance 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f was re-scheduled: Binding failed for port e4af3f09-3554-4008-bd79-56dd4a9affe7, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 694.424123] env[62569]: DEBUG nova.compute.manager [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 694.424354] env[62569]: DEBUG oslo_concurrency.lockutils [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "refresh_cache-0eff7f21-17f8-4642-bfc4-a7a15bdfb41f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.424498] env[62569]: DEBUG oslo_concurrency.lockutils [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired lock "refresh_cache-0eff7f21-17f8-4642-bfc4-a7a15bdfb41f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.424655] env[62569]: DEBUG nova.network.neutron [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 694.449538] env[62569]: DEBUG nova.network.neutron [-] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.471026] env[62569]: DEBUG nova.compute.manager [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 694.495821] env[62569]: DEBUG nova.virt.hardware [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 694.496107] env[62569]: DEBUG nova.virt.hardware [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 694.496273] env[62569]: DEBUG nova.virt.hardware [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 694.496455] env[62569]: DEBUG nova.virt.hardware [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 694.496602] env[62569]: DEBUG nova.virt.hardware [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 694.496749] env[62569]: DEBUG nova.virt.hardware [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 694.496953] env[62569]: DEBUG nova.virt.hardware [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 694.497133] env[62569]: DEBUG nova.virt.hardware [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 694.497302] env[62569]: DEBUG nova.virt.hardware [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 694.497464] env[62569]: DEBUG nova.virt.hardware [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 694.497657] env[62569]: DEBUG nova.virt.hardware [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 694.498780] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec1586aa-3701-4d9e-89d8-bc2babc81903 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.506567] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54b60592-e541-4b02-b66a-2b6cffe60c4d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.520203] env[62569]: ERROR nova.compute.manager [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fe670c41-ccb2-44ed-bf68-c487c6297401, please check neutron logs for more information. [ 694.520203] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Traceback (most recent call last): [ 694.520203] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 694.520203] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] yield resources [ 694.520203] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 694.520203] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] self.driver.spawn(context, instance, image_meta, [ 694.520203] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 694.520203] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 694.520203] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 694.520203] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] vm_ref = self.build_virtual_machine(instance, [ 694.520203] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 694.520515] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] vif_infos = vmwarevif.get_vif_info(self._session, [ 694.520515] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 694.520515] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] for vif in network_info: [ 694.520515] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 694.520515] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] return self._sync_wrapper(fn, *args, **kwargs) [ 694.520515] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 694.520515] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] self.wait() [ 694.520515] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 694.520515] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] self[:] = self._gt.wait() [ 694.520515] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 694.520515] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] return self._exit_event.wait() [ 694.520515] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 694.520515] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] current.throw(*self._exc) [ 694.520842] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.520842] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] result = function(*args, **kwargs) [ 694.520842] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.520842] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] return func(*args, **kwargs) [ 694.520842] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 694.520842] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] raise e [ 694.520842] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 694.520842] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] nwinfo = self.network_api.allocate_for_instance( [ 694.520842] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.520842] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] created_port_ids = self._update_ports_for_instance( [ 694.520842] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.520842] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] with excutils.save_and_reraise_exception(): [ 694.520842] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.521199] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] self.force_reraise() [ 694.521199] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.521199] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] raise self.value [ 694.521199] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.521199] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] updated_port = self._update_port( [ 694.521199] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.521199] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] _ensure_no_port_binding_failure(port) [ 694.521199] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.521199] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] raise exception.PortBindingFailed(port_id=port['id']) [ 694.521199] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] nova.exception.PortBindingFailed: Binding failed for port fe670c41-ccb2-44ed-bf68-c487c6297401, please check neutron logs for more information. [ 694.521199] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] [ 694.521199] env[62569]: INFO nova.compute.manager [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Terminating instance [ 694.943741] env[62569]: DEBUG nova.network.neutron [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.952230] env[62569]: INFO nova.compute.manager [-] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Took 1.02 seconds to deallocate network for instance. [ 694.954328] env[62569]: DEBUG nova.compute.claims [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 694.954511] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.024902] env[62569]: DEBUG oslo_concurrency.lockutils [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Acquiring lock "refresh_cache-bda4e4c4-b1e2-4441-8aa5-6d1b5975647c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.025126] env[62569]: DEBUG oslo_concurrency.lockutils [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Acquired lock "refresh_cache-bda4e4c4-b1e2-4441-8aa5-6d1b5975647c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.025409] env[62569]: DEBUG nova.network.neutron [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 695.027206] env[62569]: DEBUG nova.network.neutron [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.534768] env[62569]: DEBUG oslo_concurrency.lockutils [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Releasing lock "refresh_cache-0eff7f21-17f8-4642-bfc4-a7a15bdfb41f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.535032] env[62569]: DEBUG nova.compute.manager [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 695.535398] env[62569]: DEBUG nova.compute.manager [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 695.535620] env[62569]: DEBUG nova.network.neutron [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 695.558274] env[62569]: DEBUG nova.network.neutron [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.559900] env[62569]: DEBUG nova.network.neutron [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.637884] env[62569]: DEBUG nova.network.neutron [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.659859] env[62569]: DEBUG nova.compute.manager [req-8e8d32de-1fb5-4e27-955e-b9d7f93de1de req-c8b47e3f-e11f-432c-a860-2d0d15c2d9bb service nova] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Received event network-changed-fe670c41-ccb2-44ed-bf68-c487c6297401 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 695.660133] env[62569]: DEBUG nova.compute.manager [req-8e8d32de-1fb5-4e27-955e-b9d7f93de1de req-c8b47e3f-e11f-432c-a860-2d0d15c2d9bb service nova] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Refreshing instance network info cache due to event network-changed-fe670c41-ccb2-44ed-bf68-c487c6297401. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 695.660287] env[62569]: DEBUG oslo_concurrency.lockutils [req-8e8d32de-1fb5-4e27-955e-b9d7f93de1de req-c8b47e3f-e11f-432c-a860-2d0d15c2d9bb service nova] Acquiring lock "refresh_cache-bda4e4c4-b1e2-4441-8aa5-6d1b5975647c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.763544] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2052e52-cc95-4986-b15f-9b76f1c94166 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.770875] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c24f4a8-f43d-41c5-b6e6-bc94ee5255cf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.801402] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c09ff1b0-d261-432f-9216-90c8ab273053 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.807533] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d830f3-0a2b-4d52-a156-971316cac5e7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.820590] env[62569]: DEBUG nova.compute.provider_tree [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.063163] env[62569]: DEBUG nova.network.neutron [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.142574] env[62569]: DEBUG oslo_concurrency.lockutils [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Releasing lock "refresh_cache-bda4e4c4-b1e2-4441-8aa5-6d1b5975647c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.143615] env[62569]: DEBUG nova.compute.manager [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 696.143615] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 696.143615] env[62569]: DEBUG oslo_concurrency.lockutils [req-8e8d32de-1fb5-4e27-955e-b9d7f93de1de req-c8b47e3f-e11f-432c-a860-2d0d15c2d9bb service nova] Acquired lock "refresh_cache-bda4e4c4-b1e2-4441-8aa5-6d1b5975647c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.143753] env[62569]: DEBUG nova.network.neutron [req-8e8d32de-1fb5-4e27-955e-b9d7f93de1de req-c8b47e3f-e11f-432c-a860-2d0d15c2d9bb service nova] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Refreshing network info cache for port fe670c41-ccb2-44ed-bf68-c487c6297401 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 696.144702] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fdd72ede-f4b5-4895-b7e2-11feb42a43db {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.154160] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c33abf-018f-4e1e-b42c-641aaa81abee {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.175020] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bda4e4c4-b1e2-4441-8aa5-6d1b5975647c could not be found. [ 696.175232] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 696.175410] env[62569]: INFO nova.compute.manager [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 696.175632] env[62569]: DEBUG oslo.service.loopingcall [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 696.175831] env[62569]: DEBUG nova.compute.manager [-] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 696.175922] env[62569]: DEBUG nova.network.neutron [-] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 696.190788] env[62569]: DEBUG nova.network.neutron [-] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.323635] env[62569]: DEBUG nova.scheduler.client.report [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 696.567572] env[62569]: INFO nova.compute.manager [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f] Took 1.03 seconds to deallocate network for instance. [ 696.666226] env[62569]: DEBUG nova.network.neutron [req-8e8d32de-1fb5-4e27-955e-b9d7f93de1de req-c8b47e3f-e11f-432c-a860-2d0d15c2d9bb service nova] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.692895] env[62569]: DEBUG nova.network.neutron [-] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.737724] env[62569]: DEBUG nova.network.neutron [req-8e8d32de-1fb5-4e27-955e-b9d7f93de1de req-c8b47e3f-e11f-432c-a860-2d0d15c2d9bb service nova] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.831050] env[62569]: DEBUG oslo_concurrency.lockutils [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.410s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.831050] env[62569]: DEBUG nova.compute.manager [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 696.832968] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.041s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.839140] env[62569]: INFO nova.compute.claims [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 697.197631] env[62569]: INFO nova.compute.manager [-] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Took 1.02 seconds to deallocate network for instance. [ 697.200015] env[62569]: DEBUG nova.compute.claims [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 697.200211] env[62569]: DEBUG oslo_concurrency.lockutils [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.240457] env[62569]: DEBUG oslo_concurrency.lockutils [req-8e8d32de-1fb5-4e27-955e-b9d7f93de1de req-c8b47e3f-e11f-432c-a860-2d0d15c2d9bb service nova] Releasing lock "refresh_cache-bda4e4c4-b1e2-4441-8aa5-6d1b5975647c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.240683] env[62569]: DEBUG nova.compute.manager [req-8e8d32de-1fb5-4e27-955e-b9d7f93de1de req-c8b47e3f-e11f-432c-a860-2d0d15c2d9bb service nova] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Received event network-vif-deleted-fe670c41-ccb2-44ed-bf68-c487c6297401 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 697.343651] env[62569]: DEBUG nova.compute.utils [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 697.348014] env[62569]: DEBUG nova.compute.manager [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 697.348139] env[62569]: DEBUG nova.network.neutron [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 697.386849] env[62569]: DEBUG nova.policy [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '491bf8912b7b46e09c585060b6837362', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea875a4f3ccc4b7f9bd91ccbcca2226d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 697.598484] env[62569]: INFO nova.scheduler.client.report [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Deleted allocations for instance 0eff7f21-17f8-4642-bfc4-a7a15bdfb41f [ 697.637017] env[62569]: DEBUG nova.network.neutron [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Successfully created port: 16eb07c8-022c-426c-99ab-541f1820c784 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 697.851863] env[62569]: DEBUG nova.compute.manager [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 698.111283] env[62569]: DEBUG oslo_concurrency.lockutils [None req-72adf294-b139-43e2-9786-2640f3417574 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "0eff7f21-17f8-4642-bfc4-a7a15bdfb41f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 138.867s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.247999] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130218ae-5bd8-4265-9340-14ca0e4214fb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.256425] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfd2a34e-fa75-41fa-845a-65d0894dcc7c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.286742] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53e66de8-d23c-4ad7-bdaa-1475bdbb1c7d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.294955] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdcb5a64-0576-43fa-8aaa-1c7f76df5989 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.309391] env[62569]: DEBUG nova.compute.provider_tree [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.482696] env[62569]: DEBUG nova.compute.manager [req-5cc9b29b-611e-4787-87e6-0392b6d668f9 req-355da8b5-d9ff-43ee-913d-623113b958e0 service nova] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Received event network-changed-16eb07c8-022c-426c-99ab-541f1820c784 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 698.482894] env[62569]: DEBUG nova.compute.manager [req-5cc9b29b-611e-4787-87e6-0392b6d668f9 req-355da8b5-d9ff-43ee-913d-623113b958e0 service nova] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Refreshing instance network info cache due to event network-changed-16eb07c8-022c-426c-99ab-541f1820c784. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 698.483120] env[62569]: DEBUG oslo_concurrency.lockutils [req-5cc9b29b-611e-4787-87e6-0392b6d668f9 req-355da8b5-d9ff-43ee-913d-623113b958e0 service nova] Acquiring lock "refresh_cache-a930be54-d8c9-4407-9d49-f067defc65e9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.483296] env[62569]: DEBUG oslo_concurrency.lockutils [req-5cc9b29b-611e-4787-87e6-0392b6d668f9 req-355da8b5-d9ff-43ee-913d-623113b958e0 service nova] Acquired lock "refresh_cache-a930be54-d8c9-4407-9d49-f067defc65e9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.483475] env[62569]: DEBUG nova.network.neutron [req-5cc9b29b-611e-4787-87e6-0392b6d668f9 req-355da8b5-d9ff-43ee-913d-623113b958e0 service nova] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Refreshing network info cache for port 16eb07c8-022c-426c-99ab-541f1820c784 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 698.616334] env[62569]: DEBUG nova.compute.manager [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 698.639410] env[62569]: ERROR nova.compute.manager [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 16eb07c8-022c-426c-99ab-541f1820c784, please check neutron logs for more information. [ 698.639410] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 698.639410] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 698.639410] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 698.639410] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 698.639410] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 698.639410] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 698.639410] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 698.639410] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 698.639410] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 698.639410] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 698.639410] env[62569]: ERROR nova.compute.manager raise self.value [ 698.639410] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 698.639410] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 698.639410] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 698.639410] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 698.639855] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 698.639855] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 698.639855] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 16eb07c8-022c-426c-99ab-541f1820c784, please check neutron logs for more information. [ 698.639855] env[62569]: ERROR nova.compute.manager [ 698.639855] env[62569]: Traceback (most recent call last): [ 698.639855] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 698.639855] env[62569]: listener.cb(fileno) [ 698.639855] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 698.639855] env[62569]: result = function(*args, **kwargs) [ 698.639855] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 698.639855] env[62569]: return func(*args, **kwargs) [ 698.639855] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 698.639855] env[62569]: raise e [ 698.639855] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 698.639855] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 698.639855] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 698.639855] env[62569]: created_port_ids = self._update_ports_for_instance( [ 698.639855] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 698.639855] env[62569]: with excutils.save_and_reraise_exception(): [ 698.639855] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 698.639855] env[62569]: self.force_reraise() [ 698.639855] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 698.639855] env[62569]: raise self.value [ 698.639855] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 698.639855] env[62569]: updated_port = self._update_port( [ 698.639855] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 698.639855] env[62569]: _ensure_no_port_binding_failure(port) [ 698.639855] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 698.639855] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 698.640576] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 16eb07c8-022c-426c-99ab-541f1820c784, please check neutron logs for more information. [ 698.640576] env[62569]: Removing descriptor: 14 [ 698.812618] env[62569]: DEBUG nova.scheduler.client.report [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 698.864045] env[62569]: DEBUG nova.compute.manager [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 698.887657] env[62569]: DEBUG nova.virt.hardware [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 698.887797] env[62569]: DEBUG nova.virt.hardware [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 698.887973] env[62569]: DEBUG nova.virt.hardware [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 698.888362] env[62569]: DEBUG nova.virt.hardware [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 698.888362] env[62569]: DEBUG nova.virt.hardware [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 698.888470] env[62569]: DEBUG nova.virt.hardware [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 698.888661] env[62569]: DEBUG nova.virt.hardware [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 698.888818] env[62569]: DEBUG nova.virt.hardware [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 698.888978] env[62569]: DEBUG nova.virt.hardware [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 698.889154] env[62569]: DEBUG nova.virt.hardware [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 698.889335] env[62569]: DEBUG nova.virt.hardware [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 698.890219] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aabe7480-45c9-44eb-820b-484e0c13ba34 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.897986] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c8ddb14-1d2a-49d3-9ee9-cc0466a0a05e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.912018] env[62569]: ERROR nova.compute.manager [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 16eb07c8-022c-426c-99ab-541f1820c784, please check neutron logs for more information. [ 698.912018] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Traceback (most recent call last): [ 698.912018] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 698.912018] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] yield resources [ 698.912018] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 698.912018] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] self.driver.spawn(context, instance, image_meta, [ 698.912018] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 698.912018] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 698.912018] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 698.912018] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] vm_ref = self.build_virtual_machine(instance, [ 698.912018] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 698.912357] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] vif_infos = vmwarevif.get_vif_info(self._session, [ 698.912357] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 698.912357] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] for vif in network_info: [ 698.912357] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 698.912357] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] return self._sync_wrapper(fn, *args, **kwargs) [ 698.912357] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 698.912357] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] self.wait() [ 698.912357] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 698.912357] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] self[:] = self._gt.wait() [ 698.912357] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 698.912357] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] return self._exit_event.wait() [ 698.912357] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 698.912357] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] current.throw(*self._exc) [ 698.912714] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 698.912714] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] result = function(*args, **kwargs) [ 698.912714] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 698.912714] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] return func(*args, **kwargs) [ 698.912714] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 698.912714] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] raise e [ 698.912714] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 698.912714] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] nwinfo = self.network_api.allocate_for_instance( [ 698.912714] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 698.912714] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] created_port_ids = self._update_ports_for_instance( [ 698.912714] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 698.912714] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] with excutils.save_and_reraise_exception(): [ 698.912714] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 698.913044] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] self.force_reraise() [ 698.913044] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 698.913044] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] raise self.value [ 698.913044] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 698.913044] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] updated_port = self._update_port( [ 698.913044] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 698.913044] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] _ensure_no_port_binding_failure(port) [ 698.913044] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 698.913044] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] raise exception.PortBindingFailed(port_id=port['id']) [ 698.913044] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] nova.exception.PortBindingFailed: Binding failed for port 16eb07c8-022c-426c-99ab-541f1820c784, please check neutron logs for more information. [ 698.913044] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] [ 698.913044] env[62569]: INFO nova.compute.manager [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Terminating instance [ 699.002846] env[62569]: DEBUG nova.network.neutron [req-5cc9b29b-611e-4787-87e6-0392b6d668f9 req-355da8b5-d9ff-43ee-913d-623113b958e0 service nova] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 699.083967] env[62569]: DEBUG nova.network.neutron [req-5cc9b29b-611e-4787-87e6-0392b6d668f9 req-355da8b5-d9ff-43ee-913d-623113b958e0 service nova] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.141167] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.318190] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.485s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.318716] env[62569]: DEBUG nova.compute.manager [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 699.322023] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.699s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.354339] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "eb70341b-4282-4eca-b6a2-374db7c521c5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.354649] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "eb70341b-4282-4eca-b6a2-374db7c521c5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.416963] env[62569]: DEBUG oslo_concurrency.lockutils [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Acquiring lock "refresh_cache-a930be54-d8c9-4407-9d49-f067defc65e9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.587063] env[62569]: DEBUG oslo_concurrency.lockutils [req-5cc9b29b-611e-4787-87e6-0392b6d668f9 req-355da8b5-d9ff-43ee-913d-623113b958e0 service nova] Releasing lock "refresh_cache-a930be54-d8c9-4407-9d49-f067defc65e9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.587558] env[62569]: DEBUG oslo_concurrency.lockutils [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Acquired lock "refresh_cache-a930be54-d8c9-4407-9d49-f067defc65e9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.587788] env[62569]: DEBUG nova.network.neutron [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 699.827709] env[62569]: DEBUG nova.compute.utils [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 699.832358] env[62569]: DEBUG nova.compute.manager [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 699.832358] env[62569]: DEBUG nova.network.neutron [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 699.875151] env[62569]: DEBUG nova.policy [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0a9d5c0ca83b488e9bcaec31c7867b44', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '17a68974796a49359e6498789537867a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 700.108899] env[62569]: DEBUG nova.network.neutron [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.149863] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58f92a78-2654-45f2-b21a-18fc4e4ceba4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.159095] env[62569]: DEBUG nova.network.neutron [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Successfully created port: 233921b5-3801-4f9f-bf51-2c2171e8c031 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 700.161400] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a8ed2fc-18b8-471b-9165-3c780d7516b5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.193922] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f39dda-be30-43e3-9957-f228622e7f8d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.201618] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a726f5fa-d040-4d37-bfe9-90c931f879ad {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.222287] env[62569]: DEBUG nova.compute.provider_tree [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.247355] env[62569]: DEBUG nova.network.neutron [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.332547] env[62569]: DEBUG nova.compute.manager [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 700.552552] env[62569]: DEBUG nova.compute.manager [req-fee67f2e-8f44-4586-aa73-ba12c69c2ee1 req-7990a0a8-3514-4ea8-b151-11acef490f83 service nova] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Received event network-vif-deleted-16eb07c8-022c-426c-99ab-541f1820c784 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 700.724817] env[62569]: DEBUG nova.scheduler.client.report [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 700.749318] env[62569]: DEBUG oslo_concurrency.lockutils [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Releasing lock "refresh_cache-a930be54-d8c9-4407-9d49-f067defc65e9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.749840] env[62569]: DEBUG nova.compute.manager [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 700.750084] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 700.751198] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7d2f6625-3f21-4f87-b9c5-9dfb91a7ee30 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.763660] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51e8f71f-7695-495a-bee2-a001bca7b813 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.785374] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a930be54-d8c9-4407-9d49-f067defc65e9 could not be found. [ 700.785609] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 700.785815] env[62569]: INFO nova.compute.manager [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 700.786123] env[62569]: DEBUG oslo.service.loopingcall [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 700.786345] env[62569]: DEBUG nova.compute.manager [-] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 700.786441] env[62569]: DEBUG nova.network.neutron [-] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 700.803772] env[62569]: DEBUG nova.network.neutron [-] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.032963] env[62569]: ERROR nova.compute.manager [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 233921b5-3801-4f9f-bf51-2c2171e8c031, please check neutron logs for more information. [ 701.032963] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 701.032963] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 701.032963] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 701.032963] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.032963] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 701.032963] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.032963] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 701.032963] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.032963] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 701.032963] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.032963] env[62569]: ERROR nova.compute.manager raise self.value [ 701.032963] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.032963] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 701.032963] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.032963] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 701.033718] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.033718] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 701.033718] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 233921b5-3801-4f9f-bf51-2c2171e8c031, please check neutron logs for more information. [ 701.033718] env[62569]: ERROR nova.compute.manager [ 701.033718] env[62569]: Traceback (most recent call last): [ 701.033718] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 701.033718] env[62569]: listener.cb(fileno) [ 701.033718] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.033718] env[62569]: result = function(*args, **kwargs) [ 701.033718] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.033718] env[62569]: return func(*args, **kwargs) [ 701.033718] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 701.033718] env[62569]: raise e [ 701.033718] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 701.033718] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 701.033718] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.033718] env[62569]: created_port_ids = self._update_ports_for_instance( [ 701.033718] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.033718] env[62569]: with excutils.save_and_reraise_exception(): [ 701.033718] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.033718] env[62569]: self.force_reraise() [ 701.033718] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.033718] env[62569]: raise self.value [ 701.033718] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.033718] env[62569]: updated_port = self._update_port( [ 701.033718] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.033718] env[62569]: _ensure_no_port_binding_failure(port) [ 701.033718] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.033718] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 701.034659] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 233921b5-3801-4f9f-bf51-2c2171e8c031, please check neutron logs for more information. [ 701.034659] env[62569]: Removing descriptor: 14 [ 701.229439] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.907s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.230129] env[62569]: ERROR nova.compute.manager [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a9cadb21-f24f-480a-819a-10cfdb8e514b, please check neutron logs for more information. [ 701.230129] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Traceback (most recent call last): [ 701.230129] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 701.230129] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] self.driver.spawn(context, instance, image_meta, [ 701.230129] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 701.230129] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.230129] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.230129] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] vm_ref = self.build_virtual_machine(instance, [ 701.230129] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.230129] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.230129] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.230552] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] for vif in network_info: [ 701.230552] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 701.230552] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] return self._sync_wrapper(fn, *args, **kwargs) [ 701.230552] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 701.230552] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] self.wait() [ 701.230552] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 701.230552] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] self[:] = self._gt.wait() [ 701.230552] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.230552] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] return self._exit_event.wait() [ 701.230552] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 701.230552] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] current.throw(*self._exc) [ 701.230552] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.230552] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] result = function(*args, **kwargs) [ 701.230961] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.230961] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] return func(*args, **kwargs) [ 701.230961] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 701.230961] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] raise e [ 701.230961] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 701.230961] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] nwinfo = self.network_api.allocate_for_instance( [ 701.230961] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.230961] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] created_port_ids = self._update_ports_for_instance( [ 701.230961] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.230961] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] with excutils.save_and_reraise_exception(): [ 701.230961] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.230961] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] self.force_reraise() [ 701.230961] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.231391] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] raise self.value [ 701.231391] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.231391] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] updated_port = self._update_port( [ 701.231391] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.231391] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] _ensure_no_port_binding_failure(port) [ 701.231391] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.231391] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] raise exception.PortBindingFailed(port_id=port['id']) [ 701.231391] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] nova.exception.PortBindingFailed: Binding failed for port a9cadb21-f24f-480a-819a-10cfdb8e514b, please check neutron logs for more information. [ 701.231391] env[62569]: ERROR nova.compute.manager [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] [ 701.231391] env[62569]: DEBUG nova.compute.utils [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Binding failed for port a9cadb21-f24f-480a-819a-10cfdb8e514b, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 701.232105] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.352s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.236624] env[62569]: DEBUG nova.compute.manager [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Build of instance 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97 was re-scheduled: Binding failed for port a9cadb21-f24f-480a-819a-10cfdb8e514b, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 701.236756] env[62569]: DEBUG nova.compute.manager [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 701.236941] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Acquiring lock "refresh_cache-0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.237109] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Acquired lock "refresh_cache-0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.237268] env[62569]: DEBUG nova.network.neutron [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 701.306240] env[62569]: DEBUG nova.network.neutron [-] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.343891] env[62569]: DEBUG nova.compute.manager [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 701.366681] env[62569]: DEBUG nova.virt.hardware [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 701.366919] env[62569]: DEBUG nova.virt.hardware [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 701.367087] env[62569]: DEBUG nova.virt.hardware [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 701.367275] env[62569]: DEBUG nova.virt.hardware [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 701.367419] env[62569]: DEBUG nova.virt.hardware [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 701.367563] env[62569]: DEBUG nova.virt.hardware [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 701.367978] env[62569]: DEBUG nova.virt.hardware [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 701.367978] env[62569]: DEBUG nova.virt.hardware [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 701.368146] env[62569]: DEBUG nova.virt.hardware [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 701.368310] env[62569]: DEBUG nova.virt.hardware [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 701.368480] env[62569]: DEBUG nova.virt.hardware [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 701.369318] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5df699b-0da1-4fbb-8c5d-183b8b229c8f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.378029] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c78273-173a-4c85-9365-42826531b69d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.390582] env[62569]: ERROR nova.compute.manager [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 233921b5-3801-4f9f-bf51-2c2171e8c031, please check neutron logs for more information. [ 701.390582] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Traceback (most recent call last): [ 701.390582] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 701.390582] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] yield resources [ 701.390582] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 701.390582] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] self.driver.spawn(context, instance, image_meta, [ 701.390582] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 701.390582] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.390582] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.390582] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] vm_ref = self.build_virtual_machine(instance, [ 701.390582] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.390918] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.390918] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.390918] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] for vif in network_info: [ 701.390918] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 701.390918] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] return self._sync_wrapper(fn, *args, **kwargs) [ 701.390918] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 701.390918] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] self.wait() [ 701.390918] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 701.390918] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] self[:] = self._gt.wait() [ 701.390918] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.390918] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] return self._exit_event.wait() [ 701.390918] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 701.390918] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] current.throw(*self._exc) [ 701.391371] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.391371] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] result = function(*args, **kwargs) [ 701.391371] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.391371] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] return func(*args, **kwargs) [ 701.391371] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 701.391371] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] raise e [ 701.391371] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 701.391371] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] nwinfo = self.network_api.allocate_for_instance( [ 701.391371] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.391371] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] created_port_ids = self._update_ports_for_instance( [ 701.391371] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.391371] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] with excutils.save_and_reraise_exception(): [ 701.391371] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.391827] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] self.force_reraise() [ 701.391827] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.391827] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] raise self.value [ 701.391827] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.391827] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] updated_port = self._update_port( [ 701.391827] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.391827] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] _ensure_no_port_binding_failure(port) [ 701.391827] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.391827] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] raise exception.PortBindingFailed(port_id=port['id']) [ 701.391827] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] nova.exception.PortBindingFailed: Binding failed for port 233921b5-3801-4f9f-bf51-2c2171e8c031, please check neutron logs for more information. [ 701.391827] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] [ 701.391827] env[62569]: INFO nova.compute.manager [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Terminating instance [ 701.757450] env[62569]: DEBUG nova.network.neutron [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.811136] env[62569]: INFO nova.compute.manager [-] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Took 1.02 seconds to deallocate network for instance. [ 701.813297] env[62569]: DEBUG nova.compute.claims [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 701.813484] env[62569]: DEBUG oslo_concurrency.lockutils [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.831641] env[62569]: DEBUG nova.network.neutron [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.896918] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Acquiring lock "refresh_cache-34173a5f-8e4d-4646-b3e3-d537ffe752bb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.897124] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Acquired lock "refresh_cache-34173a5f-8e4d-4646-b3e3-d537ffe752bb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.897300] env[62569]: DEBUG nova.network.neutron [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 702.063058] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38ce56ac-0207-44ea-bad4-f82bd0539173 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.070581] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e63630-8023-4b38-b717-ddd815fe1820 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.101864] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e321734b-4c30-4584-892e-e6c9ebbdf66d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.109282] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1484dbfd-caee-42b7-81b3-475c2b398c53 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.122436] env[62569]: DEBUG nova.compute.provider_tree [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.333776] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Releasing lock "refresh_cache-0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.334109] env[62569]: DEBUG nova.compute.manager [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 702.334334] env[62569]: DEBUG nova.compute.manager [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 702.334541] env[62569]: DEBUG nova.network.neutron [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 702.349910] env[62569]: DEBUG nova.network.neutron [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.417378] env[62569]: DEBUG nova.network.neutron [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.517199] env[62569]: DEBUG nova.network.neutron [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.622857] env[62569]: DEBUG nova.compute.manager [req-7ff83e09-8186-47d5-a227-85bde22066c3 req-4039044e-6092-4ac0-997e-52ecc78755b8 service nova] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Received event network-changed-233921b5-3801-4f9f-bf51-2c2171e8c031 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 702.622857] env[62569]: DEBUG nova.compute.manager [req-7ff83e09-8186-47d5-a227-85bde22066c3 req-4039044e-6092-4ac0-997e-52ecc78755b8 service nova] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Refreshing instance network info cache due to event network-changed-233921b5-3801-4f9f-bf51-2c2171e8c031. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 702.622857] env[62569]: DEBUG oslo_concurrency.lockutils [req-7ff83e09-8186-47d5-a227-85bde22066c3 req-4039044e-6092-4ac0-997e-52ecc78755b8 service nova] Acquiring lock "refresh_cache-34173a5f-8e4d-4646-b3e3-d537ffe752bb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.624555] env[62569]: DEBUG nova.scheduler.client.report [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 702.853083] env[62569]: DEBUG nova.network.neutron [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.019453] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Releasing lock "refresh_cache-34173a5f-8e4d-4646-b3e3-d537ffe752bb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.019889] env[62569]: DEBUG nova.compute.manager [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 703.020133] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 703.020447] env[62569]: DEBUG oslo_concurrency.lockutils [req-7ff83e09-8186-47d5-a227-85bde22066c3 req-4039044e-6092-4ac0-997e-52ecc78755b8 service nova] Acquired lock "refresh_cache-34173a5f-8e4d-4646-b3e3-d537ffe752bb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.020643] env[62569]: DEBUG nova.network.neutron [req-7ff83e09-8186-47d5-a227-85bde22066c3 req-4039044e-6092-4ac0-997e-52ecc78755b8 service nova] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Refreshing network info cache for port 233921b5-3801-4f9f-bf51-2c2171e8c031 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 703.021687] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ac62c9ba-ae46-440b-a0fd-f2d9600f5c9b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.031538] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a03825d7-a777-46fa-837b-aa7a89802760 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.052436] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 34173a5f-8e4d-4646-b3e3-d537ffe752bb could not be found. [ 703.052638] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 703.052816] env[62569]: INFO nova.compute.manager [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Took 0.03 seconds to destroy the instance on the hypervisor. [ 703.053059] env[62569]: DEBUG oslo.service.loopingcall [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 703.053270] env[62569]: DEBUG nova.compute.manager [-] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 703.053365] env[62569]: DEBUG nova.network.neutron [-] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 703.068701] env[62569]: DEBUG nova.network.neutron [-] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.129197] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.897s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.129855] env[62569]: ERROR nova.compute.manager [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 691d17e1-efea-40ed-972b-ff3edd53f6af, please check neutron logs for more information. [ 703.129855] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Traceback (most recent call last): [ 703.129855] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 703.129855] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] self.driver.spawn(context, instance, image_meta, [ 703.129855] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 703.129855] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 703.129855] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 703.129855] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] vm_ref = self.build_virtual_machine(instance, [ 703.129855] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 703.129855] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] vif_infos = vmwarevif.get_vif_info(self._session, [ 703.129855] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 703.130294] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] for vif in network_info: [ 703.130294] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 703.130294] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] return self._sync_wrapper(fn, *args, **kwargs) [ 703.130294] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 703.130294] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] self.wait() [ 703.130294] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 703.130294] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] self[:] = self._gt.wait() [ 703.130294] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 703.130294] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] return self._exit_event.wait() [ 703.130294] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 703.130294] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] result = hub.switch() [ 703.130294] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 703.130294] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] return self.greenlet.switch() [ 703.130665] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.130665] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] result = function(*args, **kwargs) [ 703.130665] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 703.130665] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] return func(*args, **kwargs) [ 703.130665] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 703.130665] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] raise e [ 703.130665] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 703.130665] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] nwinfo = self.network_api.allocate_for_instance( [ 703.130665] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.130665] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] created_port_ids = self._update_ports_for_instance( [ 703.130665] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.130665] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] with excutils.save_and_reraise_exception(): [ 703.130665] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.131060] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] self.force_reraise() [ 703.131060] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.131060] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] raise self.value [ 703.131060] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.131060] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] updated_port = self._update_port( [ 703.131060] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.131060] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] _ensure_no_port_binding_failure(port) [ 703.131060] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.131060] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] raise exception.PortBindingFailed(port_id=port['id']) [ 703.131060] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] nova.exception.PortBindingFailed: Binding failed for port 691d17e1-efea-40ed-972b-ff3edd53f6af, please check neutron logs for more information. [ 703.131060] env[62569]: ERROR nova.compute.manager [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] [ 703.131406] env[62569]: DEBUG nova.compute.utils [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Binding failed for port 691d17e1-efea-40ed-972b-ff3edd53f6af, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 703.132155] env[62569]: DEBUG oslo_concurrency.lockutils [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.807s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.133565] env[62569]: INFO nova.compute.claims [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 703.135970] env[62569]: DEBUG nova.compute.manager [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Build of instance c3f25b97-0e1e-4e78-94f0-e8087d1b55c1 was re-scheduled: Binding failed for port 691d17e1-efea-40ed-972b-ff3edd53f6af, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 703.136397] env[62569]: DEBUG nova.compute.manager [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 703.136617] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "refresh_cache-c3f25b97-0e1e-4e78-94f0-e8087d1b55c1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.136762] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "refresh_cache-c3f25b97-0e1e-4e78-94f0-e8087d1b55c1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.136919] env[62569]: DEBUG nova.network.neutron [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 703.357972] env[62569]: INFO nova.compute.manager [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97] Took 1.02 seconds to deallocate network for instance. [ 703.542108] env[62569]: DEBUG nova.network.neutron [req-7ff83e09-8186-47d5-a227-85bde22066c3 req-4039044e-6092-4ac0-997e-52ecc78755b8 service nova] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.571432] env[62569]: DEBUG nova.network.neutron [-] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.612922] env[62569]: DEBUG nova.network.neutron [req-7ff83e09-8186-47d5-a227-85bde22066c3 req-4039044e-6092-4ac0-997e-52ecc78755b8 service nova] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.657196] env[62569]: DEBUG nova.network.neutron [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.728678] env[62569]: DEBUG nova.network.neutron [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.074072] env[62569]: INFO nova.compute.manager [-] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Took 1.02 seconds to deallocate network for instance. [ 704.078741] env[62569]: DEBUG nova.compute.claims [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 704.078949] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.117756] env[62569]: DEBUG oslo_concurrency.lockutils [req-7ff83e09-8186-47d5-a227-85bde22066c3 req-4039044e-6092-4ac0-997e-52ecc78755b8 service nova] Releasing lock "refresh_cache-34173a5f-8e4d-4646-b3e3-d537ffe752bb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.118152] env[62569]: DEBUG nova.compute.manager [req-7ff83e09-8186-47d5-a227-85bde22066c3 req-4039044e-6092-4ac0-997e-52ecc78755b8 service nova] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Received event network-vif-deleted-233921b5-3801-4f9f-bf51-2c2171e8c031 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 704.232868] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "refresh_cache-c3f25b97-0e1e-4e78-94f0-e8087d1b55c1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.232868] env[62569]: DEBUG nova.compute.manager [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 704.232868] env[62569]: DEBUG nova.compute.manager [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 704.232868] env[62569]: DEBUG nova.network.neutron [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 704.248043] env[62569]: DEBUG nova.network.neutron [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.387095] env[62569]: INFO nova.scheduler.client.report [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Deleted allocations for instance 0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97 [ 704.451305] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d92e9318-333e-4db8-8d0e-9d1411d5235e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.458909] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f686540-2e77-46e6-bb09-5f1dc6aacbea {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.487842] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-830ced46-99fd-4d60-bc28-1cdf65b00e38 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.494965] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aa7d159-68ee-486a-ba4e-a91cd980e120 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.508460] env[62569]: DEBUG nova.compute.provider_tree [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 704.750652] env[62569]: DEBUG nova.network.neutron [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.897735] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dbf03318-3818-494c-86d2-c47f300badc4 tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Lock "0f9cb1f6-e5b9-4f67-a45a-8e47e2a43a97" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.902s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.011811] env[62569]: DEBUG nova.scheduler.client.report [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 705.253864] env[62569]: INFO nova.compute.manager [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: c3f25b97-0e1e-4e78-94f0-e8087d1b55c1] Took 1.02 seconds to deallocate network for instance. [ 705.400562] env[62569]: DEBUG nova.compute.manager [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 705.517011] env[62569]: DEBUG oslo_concurrency.lockutils [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.385s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.517445] env[62569]: DEBUG nova.compute.manager [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 705.520228] env[62569]: DEBUG oslo_concurrency.lockutils [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.950s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.923852] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.025016] env[62569]: DEBUG nova.compute.utils [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 706.026431] env[62569]: DEBUG nova.compute.manager [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 706.026604] env[62569]: DEBUG nova.network.neutron [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 706.091196] env[62569]: DEBUG nova.policy [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4613deb515db4042b02af229bb05e9ca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ed0f5952cbd45eaa31512d6df93dc9e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 706.166983] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Acquiring lock "cefbca2e-609d-4954-bec6-52ffe095446f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.167297] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Lock "cefbca2e-609d-4954-bec6-52ffe095446f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.283696] env[62569]: INFO nova.scheduler.client.report [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Deleted allocations for instance c3f25b97-0e1e-4e78-94f0-e8087d1b55c1 [ 706.349533] env[62569]: DEBUG nova.network.neutron [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Successfully created port: 0273c972-0a47-491a-879c-108a4e2201e6 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 706.395189] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-822b73b7-91c5-4674-8edc-1d9dfce435f4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.402563] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b0e738-8c6f-475e-8ae7-8c4b4f26a223 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.432892] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9bd799d-431d-4eca-82ba-f0f2acdaa9bd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.439970] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6623807f-8375-4d8f-80f9-1c0149b4cfbc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.453655] env[62569]: DEBUG nova.compute.provider_tree [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.529872] env[62569]: DEBUG nova.compute.manager [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 706.580919] env[62569]: DEBUG nova.network.neutron [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Successfully created port: 8b7c2dc6-d9ff-4c01-8f1d-f0c03d52e5f5 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 706.794857] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f43bda07-3a54-4a7f-9ff4-4cf703cf9a52 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "c3f25b97-0e1e-4e78-94f0-e8087d1b55c1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.017s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.835557] env[62569]: DEBUG nova.network.neutron [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Successfully created port: 19f226f9-59ba-4afd-abd3-5dc01019bdb6 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 706.956355] env[62569]: DEBUG nova.scheduler.client.report [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 707.296079] env[62569]: DEBUG nova.compute.manager [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 707.464121] env[62569]: DEBUG oslo_concurrency.lockutils [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.942s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.464121] env[62569]: ERROR nova.compute.manager [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 99b36417-48ee-4881-a950-cfca6217ccb8, please check neutron logs for more information. [ 707.464121] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Traceback (most recent call last): [ 707.464121] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 707.464121] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] self.driver.spawn(context, instance, image_meta, [ 707.464121] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 707.464121] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] self._vmops.spawn(context, instance, image_meta, injected_files, [ 707.464121] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 707.464121] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] vm_ref = self.build_virtual_machine(instance, [ 707.464463] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 707.464463] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] vif_infos = vmwarevif.get_vif_info(self._session, [ 707.464463] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 707.464463] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] for vif in network_info: [ 707.464463] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 707.464463] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] return self._sync_wrapper(fn, *args, **kwargs) [ 707.464463] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 707.464463] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] self.wait() [ 707.464463] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 707.464463] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] self[:] = self._gt.wait() [ 707.464463] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 707.464463] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] return self._exit_event.wait() [ 707.464463] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 707.464812] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] current.throw(*self._exc) [ 707.464812] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.464812] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] result = function(*args, **kwargs) [ 707.464812] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 707.464812] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] return func(*args, **kwargs) [ 707.464812] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 707.464812] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] raise e [ 707.464812] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 707.464812] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] nwinfo = self.network_api.allocate_for_instance( [ 707.464812] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 707.464812] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] created_port_ids = self._update_ports_for_instance( [ 707.464812] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 707.464812] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] with excutils.save_and_reraise_exception(): [ 707.465170] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.465170] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] self.force_reraise() [ 707.465170] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.465170] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] raise self.value [ 707.465170] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 707.465170] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] updated_port = self._update_port( [ 707.465170] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.465170] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] _ensure_no_port_binding_failure(port) [ 707.465170] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.465170] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] raise exception.PortBindingFailed(port_id=port['id']) [ 707.465170] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] nova.exception.PortBindingFailed: Binding failed for port 99b36417-48ee-4881-a950-cfca6217ccb8, please check neutron logs for more information. [ 707.465170] env[62569]: ERROR nova.compute.manager [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] [ 707.465480] env[62569]: DEBUG nova.compute.utils [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Binding failed for port 99b36417-48ee-4881-a950-cfca6217ccb8, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 707.466182] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.230s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.472976] env[62569]: INFO nova.compute.claims [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 707.480025] env[62569]: DEBUG nova.compute.manager [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Build of instance d3274d59-6cf7-4e86-90f9-ffea49ad5342 was re-scheduled: Binding failed for port 99b36417-48ee-4881-a950-cfca6217ccb8, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 707.480025] env[62569]: DEBUG nova.compute.manager [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 707.480025] env[62569]: DEBUG oslo_concurrency.lockutils [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Acquiring lock "refresh_cache-d3274d59-6cf7-4e86-90f9-ffea49ad5342" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.480025] env[62569]: DEBUG oslo_concurrency.lockutils [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Acquired lock "refresh_cache-d3274d59-6cf7-4e86-90f9-ffea49ad5342" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.480245] env[62569]: DEBUG nova.network.neutron [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 707.540512] env[62569]: DEBUG nova.compute.manager [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 707.576830] env[62569]: DEBUG nova.virt.hardware [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 707.577332] env[62569]: DEBUG nova.virt.hardware [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 707.577793] env[62569]: DEBUG nova.virt.hardware [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 707.578082] env[62569]: DEBUG nova.virt.hardware [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 707.578405] env[62569]: DEBUG nova.virt.hardware [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 707.578698] env[62569]: DEBUG nova.virt.hardware [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 707.579094] env[62569]: DEBUG nova.virt.hardware [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 707.579538] env[62569]: DEBUG nova.virt.hardware [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 707.579875] env[62569]: DEBUG nova.virt.hardware [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 707.581104] env[62569]: DEBUG nova.virt.hardware [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 707.581463] env[62569]: DEBUG nova.virt.hardware [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 707.582791] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab1315f-43ff-41b4-b224-3bc348282ab3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.594297] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adf36b62-072f-4ceb-b23a-8c3904b891a1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.626069] env[62569]: DEBUG nova.compute.manager [req-dedddd81-9927-490f-973b-166f9ec8861b req-aedbed4c-4b1a-4ff1-af78-d5cd3740b7c4 service nova] [instance: 68218523-ae44-4332-8b20-b23559304d60] Received event network-changed-0273c972-0a47-491a-879c-108a4e2201e6 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 707.626279] env[62569]: DEBUG nova.compute.manager [req-dedddd81-9927-490f-973b-166f9ec8861b req-aedbed4c-4b1a-4ff1-af78-d5cd3740b7c4 service nova] [instance: 68218523-ae44-4332-8b20-b23559304d60] Refreshing instance network info cache due to event network-changed-0273c972-0a47-491a-879c-108a4e2201e6. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 707.626485] env[62569]: DEBUG oslo_concurrency.lockutils [req-dedddd81-9927-490f-973b-166f9ec8861b req-aedbed4c-4b1a-4ff1-af78-d5cd3740b7c4 service nova] Acquiring lock "refresh_cache-68218523-ae44-4332-8b20-b23559304d60" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.626624] env[62569]: DEBUG oslo_concurrency.lockutils [req-dedddd81-9927-490f-973b-166f9ec8861b req-aedbed4c-4b1a-4ff1-af78-d5cd3740b7c4 service nova] Acquired lock "refresh_cache-68218523-ae44-4332-8b20-b23559304d60" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.626779] env[62569]: DEBUG nova.network.neutron [req-dedddd81-9927-490f-973b-166f9ec8861b req-aedbed4c-4b1a-4ff1-af78-d5cd3740b7c4 service nova] [instance: 68218523-ae44-4332-8b20-b23559304d60] Refreshing network info cache for port 0273c972-0a47-491a-879c-108a4e2201e6 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 707.822171] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.838135] env[62569]: ERROR nova.compute.manager [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0273c972-0a47-491a-879c-108a4e2201e6, please check neutron logs for more information. [ 707.838135] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 707.838135] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 707.838135] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 707.838135] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 707.838135] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 707.838135] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 707.838135] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 707.838135] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.838135] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 707.838135] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.838135] env[62569]: ERROR nova.compute.manager raise self.value [ 707.838135] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 707.838135] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 707.838135] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.838135] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 707.838691] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.838691] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 707.838691] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0273c972-0a47-491a-879c-108a4e2201e6, please check neutron logs for more information. [ 707.838691] env[62569]: ERROR nova.compute.manager [ 707.838691] env[62569]: Traceback (most recent call last): [ 707.838691] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 707.838691] env[62569]: listener.cb(fileno) [ 707.838691] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.838691] env[62569]: result = function(*args, **kwargs) [ 707.838691] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 707.838691] env[62569]: return func(*args, **kwargs) [ 707.838691] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 707.838691] env[62569]: raise e [ 707.838691] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 707.838691] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 707.838691] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 707.838691] env[62569]: created_port_ids = self._update_ports_for_instance( [ 707.838691] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 707.838691] env[62569]: with excutils.save_and_reraise_exception(): [ 707.838691] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.838691] env[62569]: self.force_reraise() [ 707.838691] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.838691] env[62569]: raise self.value [ 707.838691] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 707.838691] env[62569]: updated_port = self._update_port( [ 707.838691] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.838691] env[62569]: _ensure_no_port_binding_failure(port) [ 707.838691] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.838691] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 707.839911] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 0273c972-0a47-491a-879c-108a4e2201e6, please check neutron logs for more information. [ 707.839911] env[62569]: Removing descriptor: 14 [ 707.839911] env[62569]: ERROR nova.compute.manager [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0273c972-0a47-491a-879c-108a4e2201e6, please check neutron logs for more information. [ 707.839911] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] Traceback (most recent call last): [ 707.839911] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 707.839911] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] yield resources [ 707.839911] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 707.839911] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] self.driver.spawn(context, instance, image_meta, [ 707.839911] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 707.839911] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] self._vmops.spawn(context, instance, image_meta, injected_files, [ 707.839911] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 707.839911] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] vm_ref = self.build_virtual_machine(instance, [ 707.840861] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 707.840861] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] vif_infos = vmwarevif.get_vif_info(self._session, [ 707.840861] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 707.840861] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] for vif in network_info: [ 707.840861] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 707.840861] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] return self._sync_wrapper(fn, *args, **kwargs) [ 707.840861] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 707.840861] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] self.wait() [ 707.840861] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 707.840861] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] self[:] = self._gt.wait() [ 707.840861] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 707.840861] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] return self._exit_event.wait() [ 707.840861] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 707.841359] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] result = hub.switch() [ 707.841359] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 707.841359] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] return self.greenlet.switch() [ 707.841359] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.841359] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] result = function(*args, **kwargs) [ 707.841359] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 707.841359] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] return func(*args, **kwargs) [ 707.841359] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 707.841359] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] raise e [ 707.841359] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 707.841359] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] nwinfo = self.network_api.allocate_for_instance( [ 707.841359] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 707.841359] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] created_port_ids = self._update_ports_for_instance( [ 707.841789] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 707.841789] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] with excutils.save_and_reraise_exception(): [ 707.841789] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.841789] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] self.force_reraise() [ 707.841789] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.841789] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] raise self.value [ 707.841789] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 707.841789] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] updated_port = self._update_port( [ 707.841789] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.841789] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] _ensure_no_port_binding_failure(port) [ 707.841789] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.841789] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] raise exception.PortBindingFailed(port_id=port['id']) [ 707.842200] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] nova.exception.PortBindingFailed: Binding failed for port 0273c972-0a47-491a-879c-108a4e2201e6, please check neutron logs for more information. [ 707.842200] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] [ 707.842200] env[62569]: INFO nova.compute.manager [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Terminating instance [ 708.002803] env[62569]: DEBUG nova.network.neutron [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.147965] env[62569]: DEBUG nova.network.neutron [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.184080] env[62569]: DEBUG nova.network.neutron [req-dedddd81-9927-490f-973b-166f9ec8861b req-aedbed4c-4b1a-4ff1-af78-d5cd3740b7c4 service nova] [instance: 68218523-ae44-4332-8b20-b23559304d60] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.319851] env[62569]: DEBUG nova.network.neutron [req-dedddd81-9927-490f-973b-166f9ec8861b req-aedbed4c-4b1a-4ff1-af78-d5cd3740b7c4 service nova] [instance: 68218523-ae44-4332-8b20-b23559304d60] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.343412] env[62569]: DEBUG oslo_concurrency.lockutils [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Acquiring lock "refresh_cache-68218523-ae44-4332-8b20-b23559304d60" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.652584] env[62569]: DEBUG oslo_concurrency.lockutils [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Releasing lock "refresh_cache-d3274d59-6cf7-4e86-90f9-ffea49ad5342" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.652894] env[62569]: DEBUG nova.compute.manager [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 708.653107] env[62569]: DEBUG nova.compute.manager [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 708.653282] env[62569]: DEBUG nova.network.neutron [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 708.672879] env[62569]: DEBUG nova.network.neutron [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.822998] env[62569]: DEBUG oslo_concurrency.lockutils [req-dedddd81-9927-490f-973b-166f9ec8861b req-aedbed4c-4b1a-4ff1-af78-d5cd3740b7c4 service nova] Releasing lock "refresh_cache-68218523-ae44-4332-8b20-b23559304d60" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.823375] env[62569]: DEBUG oslo_concurrency.lockutils [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Acquired lock "refresh_cache-68218523-ae44-4332-8b20-b23559304d60" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.823598] env[62569]: DEBUG nova.network.neutron [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 708.841272] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-793af52d-6b03-47f5-8fb6-e6e3be48f33c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.849157] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bafde03-9f5f-45a6-baff-aea68cde299f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.881030] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a42022-0cbe-4e80-9b96-4b02d11de3ba {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.887897] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1a3dd56-c78a-4b21-b7de-c6ab511f1cb2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.900738] env[62569]: DEBUG nova.compute.provider_tree [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.174700] env[62569]: DEBUG nova.network.neutron [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.361508] env[62569]: DEBUG nova.network.neutron [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 709.403708] env[62569]: DEBUG nova.scheduler.client.report [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 709.445373] env[62569]: DEBUG nova.network.neutron [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.566920] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "807d3025-d6a7-4778-a829-a61e2c7495c9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.567168] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "807d3025-d6a7-4778-a829-a61e2c7495c9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.668906] env[62569]: DEBUG nova.compute.manager [req-8e1314fb-2757-4256-8eb1-9952f66c3740 req-de532130-b62a-4101-ad79-8a06994017b4 service nova] [instance: 68218523-ae44-4332-8b20-b23559304d60] Received event network-vif-deleted-0273c972-0a47-491a-879c-108a4e2201e6 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 709.676988] env[62569]: INFO nova.compute.manager [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] [instance: d3274d59-6cf7-4e86-90f9-ffea49ad5342] Took 1.02 seconds to deallocate network for instance. [ 709.910388] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.444s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.910900] env[62569]: DEBUG nova.compute.manager [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 709.913623] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.959s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.947816] env[62569]: DEBUG oslo_concurrency.lockutils [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Releasing lock "refresh_cache-68218523-ae44-4332-8b20-b23559304d60" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.948222] env[62569]: DEBUG nova.compute.manager [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 709.948434] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 709.949332] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8291f00f-512e-48d3-8c41-f7ae4e37e013 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.958060] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-907b46bd-e66a-4f57-8473-808a3ce83350 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.979711] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 68218523-ae44-4332-8b20-b23559304d60 could not be found. [ 709.979912] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 709.980124] env[62569]: INFO nova.compute.manager [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Took 0.03 seconds to destroy the instance on the hypervisor. [ 709.980410] env[62569]: DEBUG oslo.service.loopingcall [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 709.980567] env[62569]: DEBUG nova.compute.manager [-] [instance: 68218523-ae44-4332-8b20-b23559304d60] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 709.980661] env[62569]: DEBUG nova.network.neutron [-] [instance: 68218523-ae44-4332-8b20-b23559304d60] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 710.062172] env[62569]: DEBUG nova.network.neutron [-] [instance: 68218523-ae44-4332-8b20-b23559304d60] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.417683] env[62569]: DEBUG nova.compute.utils [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 710.419322] env[62569]: DEBUG nova.compute.manager [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 710.419491] env[62569]: DEBUG nova.network.neutron [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 710.461236] env[62569]: DEBUG nova.policy [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d9d014eb68049228dbd5d588118f90c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fa4519e5b87e4816a3e5ce362059e167', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 710.713020] env[62569]: INFO nova.scheduler.client.report [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Deleted allocations for instance d3274d59-6cf7-4e86-90f9-ffea49ad5342 [ 710.758014] env[62569]: DEBUG nova.network.neutron [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Successfully created port: 52baa965-ca98-4ab1-8637-2c296f8ea753 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 710.775106] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eae669f-ad95-4e55-bf96-3e7da3748292 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.783597] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c901cbe3-3e6b-4ecd-86e3-502e80381a97 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.815186] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34127c34-f8db-480e-8b16-38e8b99b409b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.824953] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f11fcea0-c4ea-48c8-ba5e-630011c5ddb3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.837496] env[62569]: DEBUG nova.compute.provider_tree [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 710.908081] env[62569]: DEBUG nova.network.neutron [-] [instance: 68218523-ae44-4332-8b20-b23559304d60] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.920483] env[62569]: DEBUG nova.compute.manager [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 711.221991] env[62569]: DEBUG oslo_concurrency.lockutils [None req-60be867c-fb5a-4997-a719-c54e6db30014 tempest-ServersTestManualDisk-275265919 tempest-ServersTestManualDisk-275265919-project-member] Lock "d3274d59-6cf7-4e86-90f9-ffea49ad5342" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.335s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.340486] env[62569]: DEBUG nova.scheduler.client.report [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 711.410379] env[62569]: INFO nova.compute.manager [-] [instance: 68218523-ae44-4332-8b20-b23559304d60] Took 1.43 seconds to deallocate network for instance. [ 711.412762] env[62569]: DEBUG nova.compute.claims [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 711.412943] env[62569]: DEBUG oslo_concurrency.lockutils [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.725304] env[62569]: DEBUG nova.compute.manager [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 711.738576] env[62569]: DEBUG nova.compute.manager [req-ef807879-8793-40e3-a1d2-760f9afbfb56 req-a941b7c7-a795-457a-81cc-5813599ba237 service nova] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Received event network-changed-52baa965-ca98-4ab1-8637-2c296f8ea753 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 711.738727] env[62569]: DEBUG nova.compute.manager [req-ef807879-8793-40e3-a1d2-760f9afbfb56 req-a941b7c7-a795-457a-81cc-5813599ba237 service nova] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Refreshing instance network info cache due to event network-changed-52baa965-ca98-4ab1-8637-2c296f8ea753. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 711.738934] env[62569]: DEBUG oslo_concurrency.lockutils [req-ef807879-8793-40e3-a1d2-760f9afbfb56 req-a941b7c7-a795-457a-81cc-5813599ba237 service nova] Acquiring lock "refresh_cache-311e46eb-85ca-4262-a0a4-21eb073e0dc4" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.739090] env[62569]: DEBUG oslo_concurrency.lockutils [req-ef807879-8793-40e3-a1d2-760f9afbfb56 req-a941b7c7-a795-457a-81cc-5813599ba237 service nova] Acquired lock "refresh_cache-311e46eb-85ca-4262-a0a4-21eb073e0dc4" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.739253] env[62569]: DEBUG nova.network.neutron [req-ef807879-8793-40e3-a1d2-760f9afbfb56 req-a941b7c7-a795-457a-81cc-5813599ba237 service nova] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Refreshing network info cache for port 52baa965-ca98-4ab1-8637-2c296f8ea753 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 711.799896] env[62569]: ERROR nova.compute.manager [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 52baa965-ca98-4ab1-8637-2c296f8ea753, please check neutron logs for more information. [ 711.799896] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 711.799896] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 711.799896] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 711.799896] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.799896] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 711.799896] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.799896] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 711.799896] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.799896] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 711.799896] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.799896] env[62569]: ERROR nova.compute.manager raise self.value [ 711.799896] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.799896] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 711.799896] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.799896] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 711.800405] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.800405] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 711.800405] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 52baa965-ca98-4ab1-8637-2c296f8ea753, please check neutron logs for more information. [ 711.800405] env[62569]: ERROR nova.compute.manager [ 711.800405] env[62569]: Traceback (most recent call last): [ 711.800405] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 711.800405] env[62569]: listener.cb(fileno) [ 711.800405] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 711.800405] env[62569]: result = function(*args, **kwargs) [ 711.800405] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 711.800405] env[62569]: return func(*args, **kwargs) [ 711.800405] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 711.800405] env[62569]: raise e [ 711.800405] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 711.800405] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 711.800405] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.800405] env[62569]: created_port_ids = self._update_ports_for_instance( [ 711.800405] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.800405] env[62569]: with excutils.save_and_reraise_exception(): [ 711.800405] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.800405] env[62569]: self.force_reraise() [ 711.800405] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.800405] env[62569]: raise self.value [ 711.800405] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.800405] env[62569]: updated_port = self._update_port( [ 711.800405] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.800405] env[62569]: _ensure_no_port_binding_failure(port) [ 711.800405] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.800405] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 711.801208] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 52baa965-ca98-4ab1-8637-2c296f8ea753, please check neutron logs for more information. [ 711.801208] env[62569]: Removing descriptor: 17 [ 711.846796] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.933s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.847511] env[62569]: ERROR nova.compute.manager [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port decaf5e6-d463-4604-a2f9-3c72e5f71a6e, please check neutron logs for more information. [ 711.847511] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Traceback (most recent call last): [ 711.847511] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 711.847511] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] self.driver.spawn(context, instance, image_meta, [ 711.847511] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 711.847511] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 711.847511] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 711.847511] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] vm_ref = self.build_virtual_machine(instance, [ 711.847511] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 711.847511] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] vif_infos = vmwarevif.get_vif_info(self._session, [ 711.847511] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 711.847879] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] for vif in network_info: [ 711.847879] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 711.847879] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] return self._sync_wrapper(fn, *args, **kwargs) [ 711.847879] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 711.847879] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] self.wait() [ 711.847879] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 711.847879] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] self[:] = self._gt.wait() [ 711.847879] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 711.847879] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] return self._exit_event.wait() [ 711.847879] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 711.847879] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] current.throw(*self._exc) [ 711.847879] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 711.847879] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] result = function(*args, **kwargs) [ 711.848374] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 711.848374] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] return func(*args, **kwargs) [ 711.848374] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 711.848374] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] raise e [ 711.848374] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 711.848374] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] nwinfo = self.network_api.allocate_for_instance( [ 711.848374] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.848374] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] created_port_ids = self._update_ports_for_instance( [ 711.848374] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.848374] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] with excutils.save_and_reraise_exception(): [ 711.848374] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.848374] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] self.force_reraise() [ 711.848374] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.848713] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] raise self.value [ 711.848713] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.848713] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] updated_port = self._update_port( [ 711.848713] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.848713] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] _ensure_no_port_binding_failure(port) [ 711.848713] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.848713] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] raise exception.PortBindingFailed(port_id=port['id']) [ 711.848713] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] nova.exception.PortBindingFailed: Binding failed for port decaf5e6-d463-4604-a2f9-3c72e5f71a6e, please check neutron logs for more information. [ 711.848713] env[62569]: ERROR nova.compute.manager [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] [ 711.848713] env[62569]: DEBUG nova.compute.utils [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Binding failed for port decaf5e6-d463-4604-a2f9-3c72e5f71a6e, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 711.849428] env[62569]: DEBUG oslo_concurrency.lockutils [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.649s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.855972] env[62569]: DEBUG nova.compute.manager [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Build of instance 4f45ccc2-0352-4c25-9a92-aee8f051c0b5 was re-scheduled: Binding failed for port decaf5e6-d463-4604-a2f9-3c72e5f71a6e, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 711.856489] env[62569]: DEBUG nova.compute.manager [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 711.856719] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Acquiring lock "refresh_cache-4f45ccc2-0352-4c25-9a92-aee8f051c0b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.856899] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Acquired lock "refresh_cache-4f45ccc2-0352-4c25-9a92-aee8f051c0b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.857118] env[62569]: DEBUG nova.network.neutron [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 711.932346] env[62569]: DEBUG nova.compute.manager [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 711.959089] env[62569]: DEBUG nova.virt.hardware [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 711.959341] env[62569]: DEBUG nova.virt.hardware [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 711.959496] env[62569]: DEBUG nova.virt.hardware [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 711.959691] env[62569]: DEBUG nova.virt.hardware [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 711.959837] env[62569]: DEBUG nova.virt.hardware [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 711.960167] env[62569]: DEBUG nova.virt.hardware [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 711.960427] env[62569]: DEBUG nova.virt.hardware [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 711.960592] env[62569]: DEBUG nova.virt.hardware [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 711.960760] env[62569]: DEBUG nova.virt.hardware [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 711.960922] env[62569]: DEBUG nova.virt.hardware [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 711.961114] env[62569]: DEBUG nova.virt.hardware [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 711.961952] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a165e947-bd5e-4dc7-968a-cffb0f272aca {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.970647] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afc78903-e434-4f5a-8d91-f676f96928bb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.985201] env[62569]: ERROR nova.compute.manager [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 52baa965-ca98-4ab1-8637-2c296f8ea753, please check neutron logs for more information. [ 711.985201] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Traceback (most recent call last): [ 711.985201] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 711.985201] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] yield resources [ 711.985201] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 711.985201] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] self.driver.spawn(context, instance, image_meta, [ 711.985201] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 711.985201] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 711.985201] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 711.985201] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] vm_ref = self.build_virtual_machine(instance, [ 711.985201] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 711.985821] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] vif_infos = vmwarevif.get_vif_info(self._session, [ 711.985821] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 711.985821] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] for vif in network_info: [ 711.985821] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 711.985821] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] return self._sync_wrapper(fn, *args, **kwargs) [ 711.985821] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 711.985821] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] self.wait() [ 711.985821] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 711.985821] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] self[:] = self._gt.wait() [ 711.985821] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 711.985821] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] return self._exit_event.wait() [ 711.985821] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 711.985821] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] current.throw(*self._exc) [ 711.986299] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 711.986299] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] result = function(*args, **kwargs) [ 711.986299] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 711.986299] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] return func(*args, **kwargs) [ 711.986299] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 711.986299] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] raise e [ 711.986299] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 711.986299] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] nwinfo = self.network_api.allocate_for_instance( [ 711.986299] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.986299] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] created_port_ids = self._update_ports_for_instance( [ 711.986299] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.986299] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] with excutils.save_and_reraise_exception(): [ 711.986299] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.986736] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] self.force_reraise() [ 711.986736] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.986736] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] raise self.value [ 711.986736] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.986736] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] updated_port = self._update_port( [ 711.986736] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.986736] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] _ensure_no_port_binding_failure(port) [ 711.986736] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.986736] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] raise exception.PortBindingFailed(port_id=port['id']) [ 711.986736] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] nova.exception.PortBindingFailed: Binding failed for port 52baa965-ca98-4ab1-8637-2c296f8ea753, please check neutron logs for more information. [ 711.986736] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] [ 711.986736] env[62569]: INFO nova.compute.manager [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Terminating instance [ 712.254460] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.263752] env[62569]: DEBUG nova.network.neutron [req-ef807879-8793-40e3-a1d2-760f9afbfb56 req-a941b7c7-a795-457a-81cc-5813599ba237 service nova] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.351289] env[62569]: DEBUG nova.network.neutron [req-ef807879-8793-40e3-a1d2-760f9afbfb56 req-a941b7c7-a795-457a-81cc-5813599ba237 service nova] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.379594] env[62569]: DEBUG nova.network.neutron [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.465497] env[62569]: DEBUG nova.network.neutron [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.492329] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Acquiring lock "refresh_cache-311e46eb-85ca-4262-a0a4-21eb073e0dc4" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.716487] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd3a3484-96d7-4c34-b1d1-2fe5feae3460 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.725181] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1dc9f92-cd10-4b60-8f19-3ebc8b95f0f8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.758643] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4931681-2cb3-46c5-8641-af3a00f93dd5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.766473] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab7b193b-8ea6-43d8-92fb-9c0514694c1a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.780494] env[62569]: DEBUG nova.compute.provider_tree [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.852875] env[62569]: DEBUG oslo_concurrency.lockutils [req-ef807879-8793-40e3-a1d2-760f9afbfb56 req-a941b7c7-a795-457a-81cc-5813599ba237 service nova] Releasing lock "refresh_cache-311e46eb-85ca-4262-a0a4-21eb073e0dc4" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.853305] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Acquired lock "refresh_cache-311e46eb-85ca-4262-a0a4-21eb073e0dc4" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.853493] env[62569]: DEBUG nova.network.neutron [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 712.967744] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Releasing lock "refresh_cache-4f45ccc2-0352-4c25-9a92-aee8f051c0b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.968155] env[62569]: DEBUG nova.compute.manager [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 712.968397] env[62569]: DEBUG nova.compute.manager [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 712.968642] env[62569]: DEBUG nova.network.neutron [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 712.985036] env[62569]: DEBUG nova.network.neutron [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.283415] env[62569]: DEBUG nova.scheduler.client.report [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 713.376855] env[62569]: DEBUG nova.network.neutron [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.470860] env[62569]: DEBUG nova.network.neutron [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.488027] env[62569]: DEBUG nova.network.neutron [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.768427] env[62569]: DEBUG nova.compute.manager [req-11c85827-fb1d-44eb-bd65-8a3aa2b9f1c5 req-cea7c5e0-746e-4667-8c76-436ddd4db35c service nova] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Received event network-vif-deleted-52baa965-ca98-4ab1-8637-2c296f8ea753 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 713.789365] env[62569]: DEBUG oslo_concurrency.lockutils [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.940s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.789992] env[62569]: ERROR nova.compute.manager [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fe670c41-ccb2-44ed-bf68-c487c6297401, please check neutron logs for more information. [ 713.789992] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Traceback (most recent call last): [ 713.789992] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 713.789992] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] self.driver.spawn(context, instance, image_meta, [ 713.789992] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 713.789992] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 713.789992] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 713.789992] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] vm_ref = self.build_virtual_machine(instance, [ 713.789992] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 713.789992] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] vif_infos = vmwarevif.get_vif_info(self._session, [ 713.789992] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 713.790403] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] for vif in network_info: [ 713.790403] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 713.790403] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] return self._sync_wrapper(fn, *args, **kwargs) [ 713.790403] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 713.790403] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] self.wait() [ 713.790403] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 713.790403] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] self[:] = self._gt.wait() [ 713.790403] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 713.790403] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] return self._exit_event.wait() [ 713.790403] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 713.790403] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] current.throw(*self._exc) [ 713.790403] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 713.790403] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] result = function(*args, **kwargs) [ 713.790800] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 713.790800] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] return func(*args, **kwargs) [ 713.790800] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 713.790800] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] raise e [ 713.790800] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 713.790800] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] nwinfo = self.network_api.allocate_for_instance( [ 713.790800] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.790800] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] created_port_ids = self._update_ports_for_instance( [ 713.790800] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.790800] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] with excutils.save_and_reraise_exception(): [ 713.790800] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.790800] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] self.force_reraise() [ 713.790800] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.791238] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] raise self.value [ 713.791238] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.791238] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] updated_port = self._update_port( [ 713.791238] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.791238] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] _ensure_no_port_binding_failure(port) [ 713.791238] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.791238] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] raise exception.PortBindingFailed(port_id=port['id']) [ 713.791238] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] nova.exception.PortBindingFailed: Binding failed for port fe670c41-ccb2-44ed-bf68-c487c6297401, please check neutron logs for more information. [ 713.791238] env[62569]: ERROR nova.compute.manager [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] [ 713.791238] env[62569]: DEBUG nova.compute.utils [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Binding failed for port fe670c41-ccb2-44ed-bf68-c487c6297401, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 713.792633] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.651s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.793766] env[62569]: INFO nova.compute.claims [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 713.796708] env[62569]: DEBUG nova.compute.manager [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Build of instance bda4e4c4-b1e2-4441-8aa5-6d1b5975647c was re-scheduled: Binding failed for port fe670c41-ccb2-44ed-bf68-c487c6297401, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 713.797171] env[62569]: DEBUG nova.compute.manager [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 713.797389] env[62569]: DEBUG oslo_concurrency.lockutils [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Acquiring lock "refresh_cache-bda4e4c4-b1e2-4441-8aa5-6d1b5975647c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.797566] env[62569]: DEBUG oslo_concurrency.lockutils [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Acquired lock "refresh_cache-bda4e4c4-b1e2-4441-8aa5-6d1b5975647c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.797739] env[62569]: DEBUG nova.network.neutron [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 713.974202] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Releasing lock "refresh_cache-311e46eb-85ca-4262-a0a4-21eb073e0dc4" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.974600] env[62569]: DEBUG nova.compute.manager [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 713.974795] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 713.975145] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-561dcd0d-556d-4c91-80a0-70e1df5e850e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.990498] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8acaa2e9-b788-49ba-b7c7-a0fb32d17c3b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.004750] env[62569]: INFO nova.compute.manager [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 4f45ccc2-0352-4c25-9a92-aee8f051c0b5] Took 1.04 seconds to deallocate network for instance. [ 714.018746] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 311e46eb-85ca-4262-a0a4-21eb073e0dc4 could not be found. [ 714.018988] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 714.019581] env[62569]: INFO nova.compute.manager [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 714.019581] env[62569]: DEBUG oslo.service.loopingcall [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 714.019725] env[62569]: DEBUG nova.compute.manager [-] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 714.020488] env[62569]: DEBUG nova.network.neutron [-] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 714.036251] env[62569]: DEBUG nova.network.neutron [-] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.318128] env[62569]: DEBUG nova.network.neutron [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.406086] env[62569]: DEBUG nova.network.neutron [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.539216] env[62569]: DEBUG nova.network.neutron [-] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.909317] env[62569]: DEBUG oslo_concurrency.lockutils [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Releasing lock "refresh_cache-bda4e4c4-b1e2-4441-8aa5-6d1b5975647c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.909579] env[62569]: DEBUG nova.compute.manager [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 714.909725] env[62569]: DEBUG nova.compute.manager [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 714.909890] env[62569]: DEBUG nova.network.neutron [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 714.926611] env[62569]: DEBUG nova.network.neutron [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.034839] env[62569]: INFO nova.scheduler.client.report [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Deleted allocations for instance 4f45ccc2-0352-4c25-9a92-aee8f051c0b5 [ 715.045755] env[62569]: INFO nova.compute.manager [-] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Took 1.03 seconds to deallocate network for instance. [ 715.049935] env[62569]: DEBUG nova.compute.claims [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 715.050203] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.131915] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cae2ef37-1956-4321-9ce7-e4654aa078e6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.139616] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1b067ce-57a1-481b-b372-631b49c56a04 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.171412] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaeb1f9a-83d2-48bf-9f09-12909366d506 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.178959] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33ddd3e4-3e55-4679-8341-816c87762244 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.192520] env[62569]: DEBUG nova.compute.provider_tree [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.429081] env[62569]: DEBUG nova.network.neutron [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.544021] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e2364460-7356-4349-afb4-b8bba0dc1963 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Lock "4f45ccc2-0352-4c25-9a92-aee8f051c0b5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 138.989s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.700021] env[62569]: DEBUG nova.scheduler.client.report [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 715.933154] env[62569]: INFO nova.compute.manager [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: bda4e4c4-b1e2-4441-8aa5-6d1b5975647c] Took 1.02 seconds to deallocate network for instance. [ 716.045539] env[62569]: DEBUG nova.compute.manager [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 716.204000] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.412s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.204420] env[62569]: DEBUG nova.compute.manager [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 716.207027] env[62569]: DEBUG oslo_concurrency.lockutils [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.393s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.564037] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.711197] env[62569]: DEBUG nova.compute.utils [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 716.715415] env[62569]: DEBUG nova.compute.manager [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 716.715596] env[62569]: DEBUG nova.network.neutron [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 716.763512] env[62569]: DEBUG nova.policy [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd7a08521794f490bb3afc1fe589efea1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1c34b6d480e549598f1e9dae4838068a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 716.960118] env[62569]: INFO nova.scheduler.client.report [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Deleted allocations for instance bda4e4c4-b1e2-4441-8aa5-6d1b5975647c [ 717.029038] env[62569]: DEBUG nova.network.neutron [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Successfully created port: 5e1b59e6-be6d-4e60-a176-59fe81a57b29 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 717.033449] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36232510-c98a-4c03-ba65-5d273b96419d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.041228] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1b7be8d-5f40-43be-bbde-3f32b1afcf13 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.071501] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3abf23d-37d9-4474-8a6c-bc6a63bb25d8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.079128] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-254d835f-24e1-46d7-9afb-e3ce8ee048e7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.092438] env[62569]: DEBUG nova.compute.provider_tree [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.216526] env[62569]: DEBUG nova.compute.manager [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 717.472270] env[62569]: DEBUG oslo_concurrency.lockutils [None req-56243f8b-437c-430e-a5ea-fb99ed27de9a tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Lock "bda4e4c4-b1e2-4441-8aa5-6d1b5975647c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.441s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.595614] env[62569]: DEBUG nova.scheduler.client.report [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 717.713659] env[62569]: DEBUG nova.compute.manager [req-307da593-fd99-4ab1-bfb9-f7f9b15de534 req-5c131a15-7a29-4dbf-b37a-c2a95e119d2c service nova] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Received event network-changed-5e1b59e6-be6d-4e60-a176-59fe81a57b29 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 717.713659] env[62569]: DEBUG nova.compute.manager [req-307da593-fd99-4ab1-bfb9-f7f9b15de534 req-5c131a15-7a29-4dbf-b37a-c2a95e119d2c service nova] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Refreshing instance network info cache due to event network-changed-5e1b59e6-be6d-4e60-a176-59fe81a57b29. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 717.713659] env[62569]: DEBUG oslo_concurrency.lockutils [req-307da593-fd99-4ab1-bfb9-f7f9b15de534 req-5c131a15-7a29-4dbf-b37a-c2a95e119d2c service nova] Acquiring lock "refresh_cache-17f0965b-edf8-4590-9402-2654d5d19f4e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.713659] env[62569]: DEBUG oslo_concurrency.lockutils [req-307da593-fd99-4ab1-bfb9-f7f9b15de534 req-5c131a15-7a29-4dbf-b37a-c2a95e119d2c service nova] Acquired lock "refresh_cache-17f0965b-edf8-4590-9402-2654d5d19f4e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.713659] env[62569]: DEBUG nova.network.neutron [req-307da593-fd99-4ab1-bfb9-f7f9b15de534 req-5c131a15-7a29-4dbf-b37a-c2a95e119d2c service nova] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Refreshing network info cache for port 5e1b59e6-be6d-4e60-a176-59fe81a57b29 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 717.873270] env[62569]: ERROR nova.compute.manager [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5e1b59e6-be6d-4e60-a176-59fe81a57b29, please check neutron logs for more information. [ 717.873270] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 717.873270] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 717.873270] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 717.873270] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.873270] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 717.873270] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.873270] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 717.873270] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.873270] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 717.873270] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.873270] env[62569]: ERROR nova.compute.manager raise self.value [ 717.873270] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.873270] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 717.873270] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.873270] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 717.873795] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.873795] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 717.873795] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5e1b59e6-be6d-4e60-a176-59fe81a57b29, please check neutron logs for more information. [ 717.873795] env[62569]: ERROR nova.compute.manager [ 717.873795] env[62569]: Traceback (most recent call last): [ 717.873795] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 717.873795] env[62569]: listener.cb(fileno) [ 717.873795] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.873795] env[62569]: result = function(*args, **kwargs) [ 717.873795] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 717.873795] env[62569]: return func(*args, **kwargs) [ 717.873795] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 717.873795] env[62569]: raise e [ 717.873795] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 717.873795] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 717.873795] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.873795] env[62569]: created_port_ids = self._update_ports_for_instance( [ 717.873795] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.873795] env[62569]: with excutils.save_and_reraise_exception(): [ 717.873795] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.873795] env[62569]: self.force_reraise() [ 717.873795] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.873795] env[62569]: raise self.value [ 717.873795] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.873795] env[62569]: updated_port = self._update_port( [ 717.873795] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.873795] env[62569]: _ensure_no_port_binding_failure(port) [ 717.873795] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.873795] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 717.874770] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 5e1b59e6-be6d-4e60-a176-59fe81a57b29, please check neutron logs for more information. [ 717.874770] env[62569]: Removing descriptor: 17 [ 717.974806] env[62569]: DEBUG nova.compute.manager [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 718.102088] env[62569]: DEBUG oslo_concurrency.lockutils [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.895s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.102713] env[62569]: ERROR nova.compute.manager [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 16eb07c8-022c-426c-99ab-541f1820c784, please check neutron logs for more information. [ 718.102713] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Traceback (most recent call last): [ 718.102713] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 718.102713] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] self.driver.spawn(context, instance, image_meta, [ 718.102713] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 718.102713] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 718.102713] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 718.102713] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] vm_ref = self.build_virtual_machine(instance, [ 718.102713] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 718.102713] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] vif_infos = vmwarevif.get_vif_info(self._session, [ 718.102713] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 718.103097] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] for vif in network_info: [ 718.103097] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 718.103097] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] return self._sync_wrapper(fn, *args, **kwargs) [ 718.103097] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 718.103097] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] self.wait() [ 718.103097] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 718.103097] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] self[:] = self._gt.wait() [ 718.103097] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 718.103097] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] return self._exit_event.wait() [ 718.103097] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 718.103097] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] current.throw(*self._exc) [ 718.103097] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.103097] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] result = function(*args, **kwargs) [ 718.103495] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 718.103495] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] return func(*args, **kwargs) [ 718.103495] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 718.103495] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] raise e [ 718.103495] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 718.103495] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] nwinfo = self.network_api.allocate_for_instance( [ 718.103495] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.103495] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] created_port_ids = self._update_ports_for_instance( [ 718.103495] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.103495] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] with excutils.save_and_reraise_exception(): [ 718.103495] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.103495] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] self.force_reraise() [ 718.103495] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.103933] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] raise self.value [ 718.103933] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.103933] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] updated_port = self._update_port( [ 718.103933] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.103933] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] _ensure_no_port_binding_failure(port) [ 718.103933] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.103933] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] raise exception.PortBindingFailed(port_id=port['id']) [ 718.103933] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] nova.exception.PortBindingFailed: Binding failed for port 16eb07c8-022c-426c-99ab-541f1820c784, please check neutron logs for more information. [ 718.103933] env[62569]: ERROR nova.compute.manager [instance: a930be54-d8c9-4407-9d49-f067defc65e9] [ 718.103933] env[62569]: DEBUG nova.compute.utils [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Binding failed for port 16eb07c8-022c-426c-99ab-541f1820c784, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 718.104937] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.026s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.107679] env[62569]: DEBUG nova.compute.manager [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Build of instance a930be54-d8c9-4407-9d49-f067defc65e9 was re-scheduled: Binding failed for port 16eb07c8-022c-426c-99ab-541f1820c784, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 718.108095] env[62569]: DEBUG nova.compute.manager [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 718.108330] env[62569]: DEBUG oslo_concurrency.lockutils [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Acquiring lock "refresh_cache-a930be54-d8c9-4407-9d49-f067defc65e9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.108476] env[62569]: DEBUG oslo_concurrency.lockutils [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Acquired lock "refresh_cache-a930be54-d8c9-4407-9d49-f067defc65e9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.108647] env[62569]: DEBUG nova.network.neutron [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 718.230312] env[62569]: DEBUG nova.network.neutron [req-307da593-fd99-4ab1-bfb9-f7f9b15de534 req-5c131a15-7a29-4dbf-b37a-c2a95e119d2c service nova] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.232726] env[62569]: DEBUG nova.compute.manager [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 718.257374] env[62569]: DEBUG nova.virt.hardware [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 718.257616] env[62569]: DEBUG nova.virt.hardware [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 718.257773] env[62569]: DEBUG nova.virt.hardware [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 718.257976] env[62569]: DEBUG nova.virt.hardware [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 718.258137] env[62569]: DEBUG nova.virt.hardware [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 718.258313] env[62569]: DEBUG nova.virt.hardware [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 718.258591] env[62569]: DEBUG nova.virt.hardware [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 718.258770] env[62569]: DEBUG nova.virt.hardware [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 718.258985] env[62569]: DEBUG nova.virt.hardware [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 718.259112] env[62569]: DEBUG nova.virt.hardware [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 718.259289] env[62569]: DEBUG nova.virt.hardware [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 718.260178] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7e57722-1cbc-4a8a-8e90-58bea756aaa0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.268282] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1638950-248f-456d-a916-3b6131a160d5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.281925] env[62569]: ERROR nova.compute.manager [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5e1b59e6-be6d-4e60-a176-59fe81a57b29, please check neutron logs for more information. [ 718.281925] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Traceback (most recent call last): [ 718.281925] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 718.281925] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] yield resources [ 718.281925] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 718.281925] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] self.driver.spawn(context, instance, image_meta, [ 718.281925] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 718.281925] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 718.281925] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 718.281925] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] vm_ref = self.build_virtual_machine(instance, [ 718.281925] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 718.282359] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] vif_infos = vmwarevif.get_vif_info(self._session, [ 718.282359] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 718.282359] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] for vif in network_info: [ 718.282359] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 718.282359] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] return self._sync_wrapper(fn, *args, **kwargs) [ 718.282359] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 718.282359] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] self.wait() [ 718.282359] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 718.282359] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] self[:] = self._gt.wait() [ 718.282359] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 718.282359] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] return self._exit_event.wait() [ 718.282359] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 718.282359] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] current.throw(*self._exc) [ 718.282775] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.282775] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] result = function(*args, **kwargs) [ 718.282775] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 718.282775] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] return func(*args, **kwargs) [ 718.282775] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 718.282775] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] raise e [ 718.282775] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 718.282775] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] nwinfo = self.network_api.allocate_for_instance( [ 718.282775] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.282775] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] created_port_ids = self._update_ports_for_instance( [ 718.282775] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.282775] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] with excutils.save_and_reraise_exception(): [ 718.282775] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.283237] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] self.force_reraise() [ 718.283237] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.283237] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] raise self.value [ 718.283237] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.283237] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] updated_port = self._update_port( [ 718.283237] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.283237] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] _ensure_no_port_binding_failure(port) [ 718.283237] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.283237] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] raise exception.PortBindingFailed(port_id=port['id']) [ 718.283237] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] nova.exception.PortBindingFailed: Binding failed for port 5e1b59e6-be6d-4e60-a176-59fe81a57b29, please check neutron logs for more information. [ 718.283237] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] [ 718.283237] env[62569]: INFO nova.compute.manager [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Terminating instance [ 718.323922] env[62569]: DEBUG nova.network.neutron [req-307da593-fd99-4ab1-bfb9-f7f9b15de534 req-5c131a15-7a29-4dbf-b37a-c2a95e119d2c service nova] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.496065] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.628207] env[62569]: DEBUG nova.network.neutron [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.700153] env[62569]: DEBUG nova.network.neutron [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.787870] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Acquiring lock "refresh_cache-17f0965b-edf8-4590-9402-2654d5d19f4e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.826832] env[62569]: DEBUG oslo_concurrency.lockutils [req-307da593-fd99-4ab1-bfb9-f7f9b15de534 req-5c131a15-7a29-4dbf-b37a-c2a95e119d2c service nova] Releasing lock "refresh_cache-17f0965b-edf8-4590-9402-2654d5d19f4e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.827254] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Acquired lock "refresh_cache-17f0965b-edf8-4590-9402-2654d5d19f4e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.827443] env[62569]: DEBUG nova.network.neutron [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 718.895393] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16ab89b1-bc2c-402b-be81-893852f83aae {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.902536] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58cdb5a6-f0ee-4002-b176-2bf692264eff {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.932296] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f28dc3b6-2b15-437c-9b12-6584585401da {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.939414] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06467e8f-f712-4b62-827c-78c1aa0fa29e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.951352] env[62569]: DEBUG nova.compute.provider_tree [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.202186] env[62569]: DEBUG oslo_concurrency.lockutils [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Releasing lock "refresh_cache-a930be54-d8c9-4407-9d49-f067defc65e9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.202485] env[62569]: DEBUG nova.compute.manager [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 719.202485] env[62569]: DEBUG nova.compute.manager [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 719.202633] env[62569]: DEBUG nova.network.neutron [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 719.217829] env[62569]: DEBUG nova.network.neutron [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.347074] env[62569]: DEBUG nova.network.neutron [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.415237] env[62569]: DEBUG nova.network.neutron [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.454483] env[62569]: DEBUG nova.scheduler.client.report [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 719.720985] env[62569]: DEBUG nova.network.neutron [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.732730] env[62569]: DEBUG nova.compute.manager [req-950a68d9-cfb3-44a9-9e40-19358f0fe429 req-b6c686cd-a342-42db-808b-ab7dee720f23 service nova] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Received event network-vif-deleted-5e1b59e6-be6d-4e60-a176-59fe81a57b29 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 719.919087] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Releasing lock "refresh_cache-17f0965b-edf8-4590-9402-2654d5d19f4e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.919537] env[62569]: DEBUG nova.compute.manager [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 719.919727] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 719.920039] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4b262b51-1794-42c2-9fea-debbfca68934 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.929038] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fcc00d1-f0e5-4074-b562-3b5ce06eb863 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.949842] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 17f0965b-edf8-4590-9402-2654d5d19f4e could not be found. [ 719.950101] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 719.950298] env[62569]: INFO nova.compute.manager [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 719.950532] env[62569]: DEBUG oslo.service.loopingcall [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 719.950737] env[62569]: DEBUG nova.compute.manager [-] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 719.950833] env[62569]: DEBUG nova.network.neutron [-] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 719.959176] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.854s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.959690] env[62569]: ERROR nova.compute.manager [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 233921b5-3801-4f9f-bf51-2c2171e8c031, please check neutron logs for more information. [ 719.959690] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Traceback (most recent call last): [ 719.959690] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 719.959690] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] self.driver.spawn(context, instance, image_meta, [ 719.959690] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 719.959690] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 719.959690] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 719.959690] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] vm_ref = self.build_virtual_machine(instance, [ 719.959690] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 719.959690] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] vif_infos = vmwarevif.get_vif_info(self._session, [ 719.959690] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 719.960136] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] for vif in network_info: [ 719.960136] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 719.960136] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] return self._sync_wrapper(fn, *args, **kwargs) [ 719.960136] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 719.960136] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] self.wait() [ 719.960136] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 719.960136] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] self[:] = self._gt.wait() [ 719.960136] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 719.960136] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] return self._exit_event.wait() [ 719.960136] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 719.960136] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] current.throw(*self._exc) [ 719.960136] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 719.960136] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] result = function(*args, **kwargs) [ 719.960608] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 719.960608] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] return func(*args, **kwargs) [ 719.960608] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 719.960608] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] raise e [ 719.960608] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 719.960608] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] nwinfo = self.network_api.allocate_for_instance( [ 719.960608] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 719.960608] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] created_port_ids = self._update_ports_for_instance( [ 719.960608] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 719.960608] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] with excutils.save_and_reraise_exception(): [ 719.960608] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.960608] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] self.force_reraise() [ 719.960608] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.961074] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] raise self.value [ 719.961074] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 719.961074] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] updated_port = self._update_port( [ 719.961074] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.961074] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] _ensure_no_port_binding_failure(port) [ 719.961074] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.961074] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] raise exception.PortBindingFailed(port_id=port['id']) [ 719.961074] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] nova.exception.PortBindingFailed: Binding failed for port 233921b5-3801-4f9f-bf51-2c2171e8c031, please check neutron logs for more information. [ 719.961074] env[62569]: ERROR nova.compute.manager [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] [ 719.961074] env[62569]: DEBUG nova.compute.utils [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Binding failed for port 233921b5-3801-4f9f-bf51-2c2171e8c031, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 719.961483] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.040s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.962966] env[62569]: INFO nova.compute.claims [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 719.965597] env[62569]: DEBUG nova.compute.manager [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Build of instance 34173a5f-8e4d-4646-b3e3-d537ffe752bb was re-scheduled: Binding failed for port 233921b5-3801-4f9f-bf51-2c2171e8c031, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 719.965990] env[62569]: DEBUG nova.compute.manager [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 719.966224] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Acquiring lock "refresh_cache-34173a5f-8e4d-4646-b3e3-d537ffe752bb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.966370] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Acquired lock "refresh_cache-34173a5f-8e4d-4646-b3e3-d537ffe752bb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.966522] env[62569]: DEBUG nova.network.neutron [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 719.967776] env[62569]: DEBUG nova.network.neutron [-] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.223023] env[62569]: INFO nova.compute.manager [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] [instance: a930be54-d8c9-4407-9d49-f067defc65e9] Took 1.02 seconds to deallocate network for instance. [ 720.475041] env[62569]: DEBUG nova.network.neutron [-] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.488884] env[62569]: DEBUG nova.network.neutron [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.558354] env[62569]: DEBUG nova.network.neutron [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.979814] env[62569]: INFO nova.compute.manager [-] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Took 1.03 seconds to deallocate network for instance. [ 720.982618] env[62569]: DEBUG nova.compute.claims [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 720.982796] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.060526] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Releasing lock "refresh_cache-34173a5f-8e4d-4646-b3e3-d537ffe752bb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.060765] env[62569]: DEBUG nova.compute.manager [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 721.060962] env[62569]: DEBUG nova.compute.manager [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 721.061148] env[62569]: DEBUG nova.network.neutron [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 721.076584] env[62569]: DEBUG nova.network.neutron [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.239081] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 721.239352] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 721.247151] env[62569]: INFO nova.scheduler.client.report [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Deleted allocations for instance a930be54-d8c9-4407-9d49-f067defc65e9 [ 721.254364] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b1ef9d7-8309-420e-912f-c291cd367f11 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.264049] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d691387-acc2-4083-af07-ad2d3cae2032 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.300028] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da1e47a-0f69-4145-a2cd-03b13e6fb4e6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.308039] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e67e998-558f-40b9-bdea-ad1f8a04e92e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.323362] env[62569]: DEBUG nova.compute.provider_tree [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 721.582348] env[62569]: DEBUG nova.network.neutron [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.746402] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 721.746529] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Starting heal instance info cache {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 721.746641] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Rebuilding the list of instances to heal {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10312}} [ 721.759463] env[62569]: DEBUG oslo_concurrency.lockutils [None req-41200448-09c8-4afd-b962-5e0cc8ee6803 tempest-ImagesOneServerNegativeTestJSON-1718398631 tempest-ImagesOneServerNegativeTestJSON-1718398631-project-member] Lock "a930be54-d8c9-4407-9d49-f067defc65e9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.404s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.827475] env[62569]: DEBUG nova.scheduler.client.report [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 722.084725] env[62569]: INFO nova.compute.manager [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] [instance: 34173a5f-8e4d-4646-b3e3-d537ffe752bb] Took 1.02 seconds to deallocate network for instance. [ 722.250306] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 68218523-ae44-4332-8b20-b23559304d60] Skipping network cache update for instance because it is Building. {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 722.250563] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Skipping network cache update for instance because it is Building. {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 722.250599] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Skipping network cache update for instance because it is Building. {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 722.250708] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Skipping network cache update for instance because it is Building. {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 722.250831] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Didn't find any instances for network info cache update. {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10394}} [ 722.251030] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 722.251193] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 722.251335] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 722.251470] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 722.251610] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 722.251752] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 722.251876] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62569) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 722.257971] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 722.262194] env[62569]: DEBUG nova.compute.manager [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 722.334843] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.370s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.334843] env[62569]: DEBUG nova.compute.manager [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 722.336507] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.514s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.340298] env[62569]: INFO nova.compute.claims [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 722.755826] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.789325] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.839636] env[62569]: DEBUG nova.compute.utils [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 722.840258] env[62569]: DEBUG nova.compute.manager [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 722.840601] env[62569]: DEBUG nova.network.neutron [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 722.910174] env[62569]: DEBUG nova.policy [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f56e4824762e4c3b87f34d34a82c2d8f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2eb35aef726429e850178960f2ff246', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 723.122083] env[62569]: INFO nova.scheduler.client.report [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Deleted allocations for instance 34173a5f-8e4d-4646-b3e3-d537ffe752bb [ 723.344758] env[62569]: DEBUG nova.compute.manager [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 723.411643] env[62569]: DEBUG nova.network.neutron [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Successfully created port: bdb661f2-6f69-4395-a9cc-5fa2b17513b1 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 723.630696] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b0515a1d-cc06-4267-bcf1-5a08bdb42fe7 tempest-ListServerFiltersTestJSON-1107247641 tempest-ListServerFiltersTestJSON-1107247641-project-member] Lock "34173a5f-8e4d-4646-b3e3-d537ffe752bb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.206s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.724913] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e628dfad-ab23-4f2a-8fc7-53ee05854bf6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.735474] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a9e6ce2-a6d4-4291-9f65-940e193c4637 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.773943] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fccac95-3ac9-4a27-a36e-9ee92136bbf8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.781816] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9754b327-8f23-4b6a-a6d7-3118b88cecc7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.794839] env[62569]: DEBUG nova.compute.provider_tree [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 724.134237] env[62569]: DEBUG nova.compute.manager [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 724.241162] env[62569]: DEBUG nova.compute.manager [req-6c51bdc9-c373-4e26-9254-810e30ba9dec req-9a28b08a-6af1-428a-8c0e-fc4c4cd3dc36 service nova] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Received event network-changed-bdb661f2-6f69-4395-a9cc-5fa2b17513b1 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 724.241494] env[62569]: DEBUG nova.compute.manager [req-6c51bdc9-c373-4e26-9254-810e30ba9dec req-9a28b08a-6af1-428a-8c0e-fc4c4cd3dc36 service nova] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Refreshing instance network info cache due to event network-changed-bdb661f2-6f69-4395-a9cc-5fa2b17513b1. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 724.241732] env[62569]: DEBUG oslo_concurrency.lockutils [req-6c51bdc9-c373-4e26-9254-810e30ba9dec req-9a28b08a-6af1-428a-8c0e-fc4c4cd3dc36 service nova] Acquiring lock "refresh_cache-3551a2e8-2f7f-475d-9c5c-2c8e88335bc2" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.241940] env[62569]: DEBUG oslo_concurrency.lockutils [req-6c51bdc9-c373-4e26-9254-810e30ba9dec req-9a28b08a-6af1-428a-8c0e-fc4c4cd3dc36 service nova] Acquired lock "refresh_cache-3551a2e8-2f7f-475d-9c5c-2c8e88335bc2" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.242039] env[62569]: DEBUG nova.network.neutron [req-6c51bdc9-c373-4e26-9254-810e30ba9dec req-9a28b08a-6af1-428a-8c0e-fc4c4cd3dc36 service nova] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Refreshing network info cache for port bdb661f2-6f69-4395-a9cc-5fa2b17513b1 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 724.298133] env[62569]: DEBUG nova.scheduler.client.report [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 724.352690] env[62569]: ERROR nova.compute.manager [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bdb661f2-6f69-4395-a9cc-5fa2b17513b1, please check neutron logs for more information. [ 724.352690] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 724.352690] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 724.352690] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 724.352690] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.352690] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 724.352690] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.352690] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 724.352690] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.352690] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 724.352690] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.352690] env[62569]: ERROR nova.compute.manager raise self.value [ 724.352690] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.352690] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 724.352690] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.352690] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 724.353431] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.353431] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 724.353431] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bdb661f2-6f69-4395-a9cc-5fa2b17513b1, please check neutron logs for more information. [ 724.353431] env[62569]: ERROR nova.compute.manager [ 724.353431] env[62569]: Traceback (most recent call last): [ 724.353431] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 724.353431] env[62569]: listener.cb(fileno) [ 724.353431] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.353431] env[62569]: result = function(*args, **kwargs) [ 724.353431] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 724.353431] env[62569]: return func(*args, **kwargs) [ 724.353431] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 724.353431] env[62569]: raise e [ 724.353431] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 724.353431] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 724.353431] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.353431] env[62569]: created_port_ids = self._update_ports_for_instance( [ 724.353431] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.353431] env[62569]: with excutils.save_and_reraise_exception(): [ 724.353431] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.353431] env[62569]: self.force_reraise() [ 724.353431] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.353431] env[62569]: raise self.value [ 724.353431] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.353431] env[62569]: updated_port = self._update_port( [ 724.353431] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.353431] env[62569]: _ensure_no_port_binding_failure(port) [ 724.353431] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.353431] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 724.354280] env[62569]: nova.exception.PortBindingFailed: Binding failed for port bdb661f2-6f69-4395-a9cc-5fa2b17513b1, please check neutron logs for more information. [ 724.354280] env[62569]: Removing descriptor: 17 [ 724.363276] env[62569]: DEBUG nova.compute.manager [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 724.387707] env[62569]: DEBUG nova.virt.hardware [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 724.387938] env[62569]: DEBUG nova.virt.hardware [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 724.388108] env[62569]: DEBUG nova.virt.hardware [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 724.388292] env[62569]: DEBUG nova.virt.hardware [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 724.388437] env[62569]: DEBUG nova.virt.hardware [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 724.388579] env[62569]: DEBUG nova.virt.hardware [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 724.388777] env[62569]: DEBUG nova.virt.hardware [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 724.388932] env[62569]: DEBUG nova.virt.hardware [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 724.389111] env[62569]: DEBUG nova.virt.hardware [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 724.389304] env[62569]: DEBUG nova.virt.hardware [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 724.389477] env[62569]: DEBUG nova.virt.hardware [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 724.390388] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58b1f41d-49e0-498e-8532-093cbf241b37 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.398508] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb8e836-6b5b-45f9-8cc3-349c0dbd4f2e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.412457] env[62569]: ERROR nova.compute.manager [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bdb661f2-6f69-4395-a9cc-5fa2b17513b1, please check neutron logs for more information. [ 724.412457] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Traceback (most recent call last): [ 724.412457] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 724.412457] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] yield resources [ 724.412457] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 724.412457] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] self.driver.spawn(context, instance, image_meta, [ 724.412457] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 724.412457] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 724.412457] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 724.412457] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] vm_ref = self.build_virtual_machine(instance, [ 724.412457] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 724.412823] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] vif_infos = vmwarevif.get_vif_info(self._session, [ 724.412823] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 724.412823] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] for vif in network_info: [ 724.412823] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 724.412823] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] return self._sync_wrapper(fn, *args, **kwargs) [ 724.412823] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 724.412823] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] self.wait() [ 724.412823] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 724.412823] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] self[:] = self._gt.wait() [ 724.412823] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 724.412823] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] return self._exit_event.wait() [ 724.412823] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 724.412823] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] current.throw(*self._exc) [ 724.413227] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.413227] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] result = function(*args, **kwargs) [ 724.413227] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 724.413227] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] return func(*args, **kwargs) [ 724.413227] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 724.413227] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] raise e [ 724.413227] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 724.413227] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] nwinfo = self.network_api.allocate_for_instance( [ 724.413227] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.413227] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] created_port_ids = self._update_ports_for_instance( [ 724.413227] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.413227] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] with excutils.save_and_reraise_exception(): [ 724.413227] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.413594] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] self.force_reraise() [ 724.413594] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.413594] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] raise self.value [ 724.413594] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.413594] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] updated_port = self._update_port( [ 724.413594] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.413594] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] _ensure_no_port_binding_failure(port) [ 724.413594] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.413594] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] raise exception.PortBindingFailed(port_id=port['id']) [ 724.413594] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] nova.exception.PortBindingFailed: Binding failed for port bdb661f2-6f69-4395-a9cc-5fa2b17513b1, please check neutron logs for more information. [ 724.413594] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] [ 724.413594] env[62569]: INFO nova.compute.manager [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Terminating instance [ 724.659626] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.762624] env[62569]: DEBUG nova.network.neutron [req-6c51bdc9-c373-4e26-9254-810e30ba9dec req-9a28b08a-6af1-428a-8c0e-fc4c4cd3dc36 service nova] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.803468] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.467s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.804182] env[62569]: DEBUG nova.compute.manager [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 724.809101] env[62569]: DEBUG oslo_concurrency.lockutils [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.396s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.865432] env[62569]: DEBUG nova.network.neutron [req-6c51bdc9-c373-4e26-9254-810e30ba9dec req-9a28b08a-6af1-428a-8c0e-fc4c4cd3dc36 service nova] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.918620] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Acquiring lock "refresh_cache-3551a2e8-2f7f-475d-9c5c-2c8e88335bc2" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.313294] env[62569]: DEBUG nova.compute.utils [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 725.317389] env[62569]: DEBUG nova.compute.manager [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 725.317491] env[62569]: DEBUG nova.network.neutron [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 725.368554] env[62569]: DEBUG oslo_concurrency.lockutils [req-6c51bdc9-c373-4e26-9254-810e30ba9dec req-9a28b08a-6af1-428a-8c0e-fc4c4cd3dc36 service nova] Releasing lock "refresh_cache-3551a2e8-2f7f-475d-9c5c-2c8e88335bc2" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.368554] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Acquired lock "refresh_cache-3551a2e8-2f7f-475d-9c5c-2c8e88335bc2" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.368821] env[62569]: DEBUG nova.network.neutron [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 725.385405] env[62569]: DEBUG nova.policy [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aeb60cbc896e4099ab79c8d0f6b9db19', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ab20dad86a7743e1ba712bb6284231bd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 725.671254] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a469ef14-39b2-4114-af96-0c8a69087624 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.679132] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17c3e4e9-abd6-4a52-a1cb-7d934ae1ed6d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.716384] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-babc6ee1-aa9b-491b-ad85-18967b6a415e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.720146] env[62569]: DEBUG nova.network.neutron [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Successfully created port: 0a71b5f5-004f-44f5-b118-f2f26d548ffa {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 725.727022] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e2dcf8e-6e04-4364-a0bf-df0f457f935d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.740556] env[62569]: DEBUG nova.compute.provider_tree [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.820585] env[62569]: DEBUG nova.compute.manager [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 725.892510] env[62569]: DEBUG nova.network.neutron [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.011988] env[62569]: DEBUG nova.network.neutron [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.243871] env[62569]: DEBUG nova.scheduler.client.report [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 726.268088] env[62569]: DEBUG nova.compute.manager [req-26e5d769-e696-4cd8-ad56-b42e9700aafc req-15dce00f-a2f3-4bb0-93eb-407a340bd7b5 service nova] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Received event network-vif-deleted-bdb661f2-6f69-4395-a9cc-5fa2b17513b1 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 726.517017] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Releasing lock "refresh_cache-3551a2e8-2f7f-475d-9c5c-2c8e88335bc2" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.517017] env[62569]: DEBUG nova.compute.manager [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 726.517017] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 726.517017] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-91cf1747-0218-4dfd-92c7-e2bca50f4991 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.524692] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ccc81d8-4297-4701-ab9a-b49d25d7ad2b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.549817] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2 could not be found. [ 726.553018] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 726.553018] env[62569]: INFO nova.compute.manager [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 726.553018] env[62569]: DEBUG oslo.service.loopingcall [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 726.553018] env[62569]: DEBUG nova.compute.manager [-] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 726.553018] env[62569]: DEBUG nova.network.neutron [-] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 726.576824] env[62569]: DEBUG nova.network.neutron [-] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.651265] env[62569]: ERROR nova.compute.manager [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0a71b5f5-004f-44f5-b118-f2f26d548ffa, please check neutron logs for more information. [ 726.651265] env[62569]: ERROR nova.compute.manager Traceback (most recent call last): [ 726.651265] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 726.651265] env[62569]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 726.651265] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.651265] env[62569]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 726.651265] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.651265] env[62569]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 726.651265] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.651265] env[62569]: ERROR nova.compute.manager self.force_reraise() [ 726.651265] env[62569]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.651265] env[62569]: ERROR nova.compute.manager raise self.value [ 726.651265] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.651265] env[62569]: ERROR nova.compute.manager updated_port = self._update_port( [ 726.651265] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.651265] env[62569]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 726.651842] env[62569]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.651842] env[62569]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 726.651842] env[62569]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0a71b5f5-004f-44f5-b118-f2f26d548ffa, please check neutron logs for more information. [ 726.651842] env[62569]: ERROR nova.compute.manager [ 726.651842] env[62569]: Traceback (most recent call last): [ 726.651842] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 726.651842] env[62569]: listener.cb(fileno) [ 726.651842] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.651842] env[62569]: result = function(*args, **kwargs) [ 726.651842] env[62569]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.651842] env[62569]: return func(*args, **kwargs) [ 726.651842] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 726.651842] env[62569]: raise e [ 726.651842] env[62569]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 726.651842] env[62569]: nwinfo = self.network_api.allocate_for_instance( [ 726.651842] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.651842] env[62569]: created_port_ids = self._update_ports_for_instance( [ 726.651842] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.651842] env[62569]: with excutils.save_and_reraise_exception(): [ 726.651842] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.651842] env[62569]: self.force_reraise() [ 726.651842] env[62569]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.651842] env[62569]: raise self.value [ 726.651842] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.651842] env[62569]: updated_port = self._update_port( [ 726.651842] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.651842] env[62569]: _ensure_no_port_binding_failure(port) [ 726.651842] env[62569]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.651842] env[62569]: raise exception.PortBindingFailed(port_id=port['id']) [ 726.652785] env[62569]: nova.exception.PortBindingFailed: Binding failed for port 0a71b5f5-004f-44f5-b118-f2f26d548ffa, please check neutron logs for more information. [ 726.652785] env[62569]: Removing descriptor: 17 [ 726.750532] env[62569]: DEBUG oslo_concurrency.lockutils [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.942s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.751195] env[62569]: ERROR nova.compute.manager [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0273c972-0a47-491a-879c-108a4e2201e6, please check neutron logs for more information. [ 726.751195] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] Traceback (most recent call last): [ 726.751195] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 726.751195] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] self.driver.spawn(context, instance, image_meta, [ 726.751195] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 726.751195] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] self._vmops.spawn(context, instance, image_meta, injected_files, [ 726.751195] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 726.751195] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] vm_ref = self.build_virtual_machine(instance, [ 726.751195] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 726.751195] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] vif_infos = vmwarevif.get_vif_info(self._session, [ 726.751195] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 726.751644] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] for vif in network_info: [ 726.751644] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 726.751644] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] return self._sync_wrapper(fn, *args, **kwargs) [ 726.751644] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 726.751644] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] self.wait() [ 726.751644] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 726.751644] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] self[:] = self._gt.wait() [ 726.751644] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 726.751644] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] return self._exit_event.wait() [ 726.751644] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 726.751644] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] result = hub.switch() [ 726.751644] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 726.751644] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] return self.greenlet.switch() [ 726.752108] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.752108] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] result = function(*args, **kwargs) [ 726.752108] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.752108] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] return func(*args, **kwargs) [ 726.752108] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 726.752108] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] raise e [ 726.752108] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 726.752108] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] nwinfo = self.network_api.allocate_for_instance( [ 726.752108] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.752108] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] created_port_ids = self._update_ports_for_instance( [ 726.752108] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.752108] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] with excutils.save_and_reraise_exception(): [ 726.752108] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.752614] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] self.force_reraise() [ 726.752614] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.752614] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] raise self.value [ 726.752614] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.752614] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] updated_port = self._update_port( [ 726.752614] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.752614] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] _ensure_no_port_binding_failure(port) [ 726.752614] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.752614] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] raise exception.PortBindingFailed(port_id=port['id']) [ 726.752614] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] nova.exception.PortBindingFailed: Binding failed for port 0273c972-0a47-491a-879c-108a4e2201e6, please check neutron logs for more information. [ 726.752614] env[62569]: ERROR nova.compute.manager [instance: 68218523-ae44-4332-8b20-b23559304d60] [ 726.753110] env[62569]: DEBUG nova.compute.utils [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Binding failed for port 0273c972-0a47-491a-879c-108a4e2201e6, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 726.753184] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.499s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.754610] env[62569]: INFO nova.compute.claims [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 726.757849] env[62569]: DEBUG nova.compute.manager [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Build of instance 68218523-ae44-4332-8b20-b23559304d60 was re-scheduled: Binding failed for port 0273c972-0a47-491a-879c-108a4e2201e6, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 726.758303] env[62569]: DEBUG nova.compute.manager [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 726.758520] env[62569]: DEBUG oslo_concurrency.lockutils [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Acquiring lock "refresh_cache-68218523-ae44-4332-8b20-b23559304d60" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.758665] env[62569]: DEBUG oslo_concurrency.lockutils [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Acquired lock "refresh_cache-68218523-ae44-4332-8b20-b23559304d60" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.758822] env[62569]: DEBUG nova.network.neutron [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 726.832816] env[62569]: DEBUG nova.compute.manager [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 726.861827] env[62569]: DEBUG nova.virt.hardware [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 726.862090] env[62569]: DEBUG nova.virt.hardware [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 726.862253] env[62569]: DEBUG nova.virt.hardware [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 726.862443] env[62569]: DEBUG nova.virt.hardware [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 726.862591] env[62569]: DEBUG nova.virt.hardware [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 726.862738] env[62569]: DEBUG nova.virt.hardware [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 726.862945] env[62569]: DEBUG nova.virt.hardware [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 726.863570] env[62569]: DEBUG nova.virt.hardware [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 726.863873] env[62569]: DEBUG nova.virt.hardware [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 726.864081] env[62569]: DEBUG nova.virt.hardware [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 726.864555] env[62569]: DEBUG nova.virt.hardware [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 726.865405] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c42c65a3-20d6-430a-91c8-d094724d429a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.878562] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f26d8a-f81f-4972-86d0-aa97c3cce42c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.892952] env[62569]: ERROR nova.compute.manager [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0a71b5f5-004f-44f5-b118-f2f26d548ffa, please check neutron logs for more information. [ 726.892952] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Traceback (most recent call last): [ 726.892952] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 726.892952] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] yield resources [ 726.892952] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 726.892952] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] self.driver.spawn(context, instance, image_meta, [ 726.892952] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 726.892952] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] self._vmops.spawn(context, instance, image_meta, injected_files, [ 726.892952] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 726.892952] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] vm_ref = self.build_virtual_machine(instance, [ 726.892952] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 726.893400] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] vif_infos = vmwarevif.get_vif_info(self._session, [ 726.893400] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 726.893400] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] for vif in network_info: [ 726.893400] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 726.893400] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] return self._sync_wrapper(fn, *args, **kwargs) [ 726.893400] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 726.893400] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] self.wait() [ 726.893400] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 726.893400] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] self[:] = self._gt.wait() [ 726.893400] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 726.893400] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] return self._exit_event.wait() [ 726.893400] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 726.893400] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] current.throw(*self._exc) [ 726.893752] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.893752] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] result = function(*args, **kwargs) [ 726.893752] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.893752] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] return func(*args, **kwargs) [ 726.893752] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 726.893752] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] raise e [ 726.893752] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 726.893752] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] nwinfo = self.network_api.allocate_for_instance( [ 726.893752] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.893752] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] created_port_ids = self._update_ports_for_instance( [ 726.893752] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.893752] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] with excutils.save_and_reraise_exception(): [ 726.893752] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.894136] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] self.force_reraise() [ 726.894136] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.894136] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] raise self.value [ 726.894136] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.894136] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] updated_port = self._update_port( [ 726.894136] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.894136] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] _ensure_no_port_binding_failure(port) [ 726.894136] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.894136] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] raise exception.PortBindingFailed(port_id=port['id']) [ 726.894136] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] nova.exception.PortBindingFailed: Binding failed for port 0a71b5f5-004f-44f5-b118-f2f26d548ffa, please check neutron logs for more information. [ 726.894136] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] [ 726.894136] env[62569]: INFO nova.compute.manager [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Terminating instance [ 727.078879] env[62569]: DEBUG nova.network.neutron [-] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.281997] env[62569]: DEBUG nova.network.neutron [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.397753] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Acquiring lock "refresh_cache-0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.397939] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Acquired lock "refresh_cache-0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.398167] env[62569]: DEBUG nova.network.neutron [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 727.551512] env[62569]: DEBUG nova.network.neutron [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.581387] env[62569]: INFO nova.compute.manager [-] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Took 1.03 seconds to deallocate network for instance. [ 727.583607] env[62569]: DEBUG nova.compute.claims [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 727.583784] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.923164] env[62569]: DEBUG nova.network.neutron [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.042511] env[62569]: DEBUG nova.network.neutron [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.054060] env[62569]: DEBUG oslo_concurrency.lockutils [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Releasing lock "refresh_cache-68218523-ae44-4332-8b20-b23559304d60" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.054493] env[62569]: DEBUG nova.compute.manager [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 728.054493] env[62569]: DEBUG nova.compute.manager [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 728.054669] env[62569]: DEBUG nova.network.neutron [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 728.081028] env[62569]: DEBUG nova.network.neutron [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.133533] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b605542-53ec-444d-b2ef-f4b70b98bf65 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.141490] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46df014d-ec67-4908-a328-913c201d9947 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.176391] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-673da178-56d7-4284-b3d6-7cb03f1c3e5b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.184009] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ce145bc-7c41-405e-aba3-087f9e146d58 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.201142] env[62569]: DEBUG nova.compute.provider_tree [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 728.392163] env[62569]: DEBUG nova.compute.manager [req-0aafc664-68a5-4f45-abf0-6bf593c26417 req-e4e44380-85ed-4332-b6a0-426d7078348f service nova] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Received event network-changed-0a71b5f5-004f-44f5-b118-f2f26d548ffa {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 728.392402] env[62569]: DEBUG nova.compute.manager [req-0aafc664-68a5-4f45-abf0-6bf593c26417 req-e4e44380-85ed-4332-b6a0-426d7078348f service nova] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Refreshing instance network info cache due to event network-changed-0a71b5f5-004f-44f5-b118-f2f26d548ffa. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 728.392587] env[62569]: DEBUG oslo_concurrency.lockutils [req-0aafc664-68a5-4f45-abf0-6bf593c26417 req-e4e44380-85ed-4332-b6a0-426d7078348f service nova] Acquiring lock "refresh_cache-0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.545950] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Releasing lock "refresh_cache-0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.546612] env[62569]: DEBUG nova.compute.manager [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 728.546932] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 728.547364] env[62569]: DEBUG oslo_concurrency.lockutils [req-0aafc664-68a5-4f45-abf0-6bf593c26417 req-e4e44380-85ed-4332-b6a0-426d7078348f service nova] Acquired lock "refresh_cache-0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.547649] env[62569]: DEBUG nova.network.neutron [req-0aafc664-68a5-4f45-abf0-6bf593c26417 req-e4e44380-85ed-4332-b6a0-426d7078348f service nova] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Refreshing network info cache for port 0a71b5f5-004f-44f5-b118-f2f26d548ffa {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 728.555611] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e82d7a20-c5b0-4276-9633-fab4f7a556a3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.563956] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d32bb407-af5c-4928-8909-80913192ab08 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.585109] env[62569]: DEBUG nova.network.neutron [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.587023] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda could not be found. [ 728.587023] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 728.587023] env[62569]: INFO nova.compute.manager [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Took 0.04 seconds to destroy the instance on the hypervisor. [ 728.587183] env[62569]: DEBUG oslo.service.loopingcall [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 728.587657] env[62569]: DEBUG nova.compute.manager [-] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 728.587757] env[62569]: DEBUG nova.network.neutron [-] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 728.603655] env[62569]: DEBUG nova.network.neutron [-] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.705471] env[62569]: DEBUG nova.scheduler.client.report [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 729.074090] env[62569]: DEBUG nova.network.neutron [req-0aafc664-68a5-4f45-abf0-6bf593c26417 req-e4e44380-85ed-4332-b6a0-426d7078348f service nova] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.089377] env[62569]: INFO nova.compute.manager [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 68218523-ae44-4332-8b20-b23559304d60] Took 1.03 seconds to deallocate network for instance. [ 729.107605] env[62569]: DEBUG nova.network.neutron [-] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.157603] env[62569]: DEBUG nova.network.neutron [req-0aafc664-68a5-4f45-abf0-6bf593c26417 req-e4e44380-85ed-4332-b6a0-426d7078348f service nova] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.214415] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.461s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.214952] env[62569]: DEBUG nova.compute.manager [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 729.220741] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.171s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.609207] env[62569]: INFO nova.compute.manager [-] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Took 1.02 seconds to deallocate network for instance. [ 729.616653] env[62569]: DEBUG nova.compute.claims [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Aborting claim: {{(pid=62569) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 729.616846] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.661032] env[62569]: DEBUG oslo_concurrency.lockutils [req-0aafc664-68a5-4f45-abf0-6bf593c26417 req-e4e44380-85ed-4332-b6a0-426d7078348f service nova] Releasing lock "refresh_cache-0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.661340] env[62569]: DEBUG nova.compute.manager [req-0aafc664-68a5-4f45-abf0-6bf593c26417 req-e4e44380-85ed-4332-b6a0-426d7078348f service nova] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Received event network-vif-deleted-0a71b5f5-004f-44f5-b118-f2f26d548ffa {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 729.730148] env[62569]: DEBUG nova.compute.utils [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 729.732533] env[62569]: DEBUG nova.compute.manager [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 729.732533] env[62569]: DEBUG nova.network.neutron [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 729.782576] env[62569]: DEBUG nova.policy [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '73c4506682ef43678bdd5a15a7d21a20', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '074644d447544fb794da91d2ce2b0c6d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 730.056614] env[62569]: DEBUG nova.network.neutron [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Successfully created port: eb2d33a3-7bed-46d7-9656-90aa7577e5e2 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 730.074405] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40f1d796-cb95-4fd0-b072-e1d745a9a6b5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.081612] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a55c49b1-95f9-4216-9199-13a3eeaf6ce4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.112902] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-113931f7-8ac3-4090-baa3-c987e09a7819 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.119727] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac12450e-dc6e-4695-88f8-a673735ef8ff {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.135172] env[62569]: DEBUG nova.compute.provider_tree [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.136962] env[62569]: INFO nova.scheduler.client.report [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Deleted allocations for instance 68218523-ae44-4332-8b20-b23559304d60 [ 730.235070] env[62569]: DEBUG nova.compute.manager [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 730.644038] env[62569]: DEBUG nova.scheduler.client.report [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 730.647863] env[62569]: DEBUG oslo_concurrency.lockutils [None req-01baed6c-5483-44ce-bed4-1576bb77a7d1 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Lock "68218523-ae44-4332-8b20-b23559304d60" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 149.499s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.150151] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.929s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.150782] env[62569]: ERROR nova.compute.manager [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 52baa965-ca98-4ab1-8637-2c296f8ea753, please check neutron logs for more information. [ 731.150782] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Traceback (most recent call last): [ 731.150782] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 731.150782] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] self.driver.spawn(context, instance, image_meta, [ 731.150782] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 731.150782] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 731.150782] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 731.150782] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] vm_ref = self.build_virtual_machine(instance, [ 731.150782] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 731.150782] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] vif_infos = vmwarevif.get_vif_info(self._session, [ 731.150782] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 731.151214] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] for vif in network_info: [ 731.151214] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 731.151214] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] return self._sync_wrapper(fn, *args, **kwargs) [ 731.151214] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 731.151214] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] self.wait() [ 731.151214] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 731.151214] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] self[:] = self._gt.wait() [ 731.151214] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 731.151214] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] return self._exit_event.wait() [ 731.151214] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 731.151214] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] current.throw(*self._exc) [ 731.151214] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 731.151214] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] result = function(*args, **kwargs) [ 731.151650] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 731.151650] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] return func(*args, **kwargs) [ 731.151650] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 731.151650] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] raise e [ 731.151650] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 731.151650] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] nwinfo = self.network_api.allocate_for_instance( [ 731.151650] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 731.151650] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] created_port_ids = self._update_ports_for_instance( [ 731.151650] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 731.151650] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] with excutils.save_and_reraise_exception(): [ 731.151650] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.151650] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] self.force_reraise() [ 731.151650] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.152157] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] raise self.value [ 731.152157] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 731.152157] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] updated_port = self._update_port( [ 731.152157] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.152157] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] _ensure_no_port_binding_failure(port) [ 731.152157] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.152157] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] raise exception.PortBindingFailed(port_id=port['id']) [ 731.152157] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] nova.exception.PortBindingFailed: Binding failed for port 52baa965-ca98-4ab1-8637-2c296f8ea753, please check neutron logs for more information. [ 731.152157] env[62569]: ERROR nova.compute.manager [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] [ 731.152157] env[62569]: DEBUG nova.compute.utils [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Binding failed for port 52baa965-ca98-4ab1-8637-2c296f8ea753, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 731.152851] env[62569]: DEBUG nova.compute.manager [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 731.155135] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.591s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.156525] env[62569]: INFO nova.compute.claims [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 731.159165] env[62569]: DEBUG nova.compute.manager [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Build of instance 311e46eb-85ca-4262-a0a4-21eb073e0dc4 was re-scheduled: Binding failed for port 52baa965-ca98-4ab1-8637-2c296f8ea753, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 731.159592] env[62569]: DEBUG nova.compute.manager [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 731.159813] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Acquiring lock "refresh_cache-311e46eb-85ca-4262-a0a4-21eb073e0dc4" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.159954] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Acquired lock "refresh_cache-311e46eb-85ca-4262-a0a4-21eb073e0dc4" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.160126] env[62569]: DEBUG nova.network.neutron [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 731.243985] env[62569]: DEBUG nova.compute.manager [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 731.268992] env[62569]: DEBUG nova.virt.hardware [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 731.269262] env[62569]: DEBUG nova.virt.hardware [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 731.269422] env[62569]: DEBUG nova.virt.hardware [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 731.269863] env[62569]: DEBUG nova.virt.hardware [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 731.270137] env[62569]: DEBUG nova.virt.hardware [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 731.270295] env[62569]: DEBUG nova.virt.hardware [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 731.270514] env[62569]: DEBUG nova.virt.hardware [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 731.270675] env[62569]: DEBUG nova.virt.hardware [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 731.270843] env[62569]: DEBUG nova.virt.hardware [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 731.271022] env[62569]: DEBUG nova.virt.hardware [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 731.271207] env[62569]: DEBUG nova.virt.hardware [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 731.272084] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46e0d9cb-95d0-4555-bd05-20b3a13f0c69 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.280230] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7d570d5-5306-4c3b-9b85-3c3ca148bc1e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.563377] env[62569]: DEBUG nova.compute.manager [req-79534df6-9791-4c2e-a547-6e99beedb8a3 req-5deff81e-317b-488d-9676-1eada368a5a4 service nova] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Received event network-vif-plugged-eb2d33a3-7bed-46d7-9656-90aa7577e5e2 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 731.563652] env[62569]: DEBUG oslo_concurrency.lockutils [req-79534df6-9791-4c2e-a547-6e99beedb8a3 req-5deff81e-317b-488d-9676-1eada368a5a4 service nova] Acquiring lock "93c8445d-689c-4f77-836f-95eafafbb6a2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.563869] env[62569]: DEBUG oslo_concurrency.lockutils [req-79534df6-9791-4c2e-a547-6e99beedb8a3 req-5deff81e-317b-488d-9676-1eada368a5a4 service nova] Lock "93c8445d-689c-4f77-836f-95eafafbb6a2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.564429] env[62569]: DEBUG oslo_concurrency.lockutils [req-79534df6-9791-4c2e-a547-6e99beedb8a3 req-5deff81e-317b-488d-9676-1eada368a5a4 service nova] Lock "93c8445d-689c-4f77-836f-95eafafbb6a2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.564767] env[62569]: DEBUG nova.compute.manager [req-79534df6-9791-4c2e-a547-6e99beedb8a3 req-5deff81e-317b-488d-9676-1eada368a5a4 service nova] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] No waiting events found dispatching network-vif-plugged-eb2d33a3-7bed-46d7-9656-90aa7577e5e2 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 731.564853] env[62569]: WARNING nova.compute.manager [req-79534df6-9791-4c2e-a547-6e99beedb8a3 req-5deff81e-317b-488d-9676-1eada368a5a4 service nova] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Received unexpected event network-vif-plugged-eb2d33a3-7bed-46d7-9656-90aa7577e5e2 for instance with vm_state building and task_state spawning. [ 731.665230] env[62569]: DEBUG nova.network.neutron [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Successfully updated port: eb2d33a3-7bed-46d7-9656-90aa7577e5e2 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 731.695688] env[62569]: DEBUG nova.network.neutron [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.698931] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.772583] env[62569]: DEBUG nova.network.neutron [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.178815] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Acquiring lock "refresh_cache-93c8445d-689c-4f77-836f-95eafafbb6a2" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.178985] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Acquired lock "refresh_cache-93c8445d-689c-4f77-836f-95eafafbb6a2" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.180393] env[62569]: DEBUG nova.network.neutron [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 732.276460] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Releasing lock "refresh_cache-311e46eb-85ca-4262-a0a4-21eb073e0dc4" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.276698] env[62569]: DEBUG nova.compute.manager [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 732.276880] env[62569]: DEBUG nova.compute.manager [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 732.277064] env[62569]: DEBUG nova.network.neutron [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 732.293305] env[62569]: DEBUG nova.network.neutron [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.490274] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b86ac99d-de07-458e-be59-03726e8d2978 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.503793] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65b603b9-9d5d-49dd-9a5b-97ec93ebd04b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.539355] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96f77a25-1863-4a70-b4b3-716d9cc08e2b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.549204] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-213034f0-0d19-4f0d-938a-91539ac77c2f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.566287] env[62569]: DEBUG nova.compute.provider_tree [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.712757] env[62569]: DEBUG nova.network.neutron [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.795697] env[62569]: DEBUG nova.network.neutron [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.911600] env[62569]: DEBUG nova.network.neutron [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Updating instance_info_cache with network_info: [{"id": "eb2d33a3-7bed-46d7-9656-90aa7577e5e2", "address": "fa:16:3e:9e:8d:2c", "network": {"id": "6f2a7189-6586-4180-86d5-ec1324a82e25", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-881682408-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "074644d447544fb794da91d2ce2b0c6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61a172ee-af3f-473e-b12a-3fee5bf39c8d", "external-id": "nsx-vlan-transportzone-997", "segmentation_id": 997, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb2d33a3-7b", "ovs_interfaceid": "eb2d33a3-7bed-46d7-9656-90aa7577e5e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.068595] env[62569]: DEBUG nova.scheduler.client.report [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 733.299372] env[62569]: INFO nova.compute.manager [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] [instance: 311e46eb-85ca-4262-a0a4-21eb073e0dc4] Took 1.02 seconds to deallocate network for instance. [ 733.323476] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Acquiring lock "390bcf25-689d-46ad-bffb-3670c3729397" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.323702] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Lock "390bcf25-689d-46ad-bffb-3670c3729397" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.413706] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Releasing lock "refresh_cache-93c8445d-689c-4f77-836f-95eafafbb6a2" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.414055] env[62569]: DEBUG nova.compute.manager [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Instance network_info: |[{"id": "eb2d33a3-7bed-46d7-9656-90aa7577e5e2", "address": "fa:16:3e:9e:8d:2c", "network": {"id": "6f2a7189-6586-4180-86d5-ec1324a82e25", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-881682408-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "074644d447544fb794da91d2ce2b0c6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61a172ee-af3f-473e-b12a-3fee5bf39c8d", "external-id": "nsx-vlan-transportzone-997", "segmentation_id": 997, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb2d33a3-7b", "ovs_interfaceid": "eb2d33a3-7bed-46d7-9656-90aa7577e5e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 733.414577] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9e:8d:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '61a172ee-af3f-473e-b12a-3fee5bf39c8d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eb2d33a3-7bed-46d7-9656-90aa7577e5e2', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 733.423041] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Creating folder: Project (074644d447544fb794da91d2ce2b0c6d). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 733.423041] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b7622de3-750a-43ce-af07-96937efc2e55 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.434295] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Created folder: Project (074644d447544fb794da91d2ce2b0c6d) in parent group-v269330. [ 733.434487] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Creating folder: Instances. Parent ref: group-v269350. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 733.434719] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-355a5b42-8656-48fe-bb03-bec7eae4b513 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.443312] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Created folder: Instances in parent group-v269350. [ 733.443641] env[62569]: DEBUG oslo.service.loopingcall [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 733.443912] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 733.444229] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ae884ec5-6161-44e4-a93d-fdf16b3e1c43 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.462349] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 733.462349] env[62569]: value = "task-1249869" [ 733.462349] env[62569]: _type = "Task" [ 733.462349] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.469260] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249869, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.573281] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.418s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.573812] env[62569]: DEBUG nova.compute.manager [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 733.576498] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.080s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.579246] env[62569]: INFO nova.compute.claims [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 733.587284] env[62569]: DEBUG nova.compute.manager [req-d4497ce9-6bce-4cc5-972a-b0a56c99215e req-e536d47e-bf06-4dac-900f-65dd373f835f service nova] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Received event network-changed-eb2d33a3-7bed-46d7-9656-90aa7577e5e2 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 733.587464] env[62569]: DEBUG nova.compute.manager [req-d4497ce9-6bce-4cc5-972a-b0a56c99215e req-e536d47e-bf06-4dac-900f-65dd373f835f service nova] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Refreshing instance network info cache due to event network-changed-eb2d33a3-7bed-46d7-9656-90aa7577e5e2. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 733.587636] env[62569]: DEBUG oslo_concurrency.lockutils [req-d4497ce9-6bce-4cc5-972a-b0a56c99215e req-e536d47e-bf06-4dac-900f-65dd373f835f service nova] Acquiring lock "refresh_cache-93c8445d-689c-4f77-836f-95eafafbb6a2" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.587814] env[62569]: DEBUG oslo_concurrency.lockutils [req-d4497ce9-6bce-4cc5-972a-b0a56c99215e req-e536d47e-bf06-4dac-900f-65dd373f835f service nova] Acquired lock "refresh_cache-93c8445d-689c-4f77-836f-95eafafbb6a2" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.587966] env[62569]: DEBUG nova.network.neutron [req-d4497ce9-6bce-4cc5-972a-b0a56c99215e req-e536d47e-bf06-4dac-900f-65dd373f835f service nova] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Refreshing network info cache for port eb2d33a3-7bed-46d7-9656-90aa7577e5e2 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 733.976570] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249869, 'name': CreateVM_Task, 'duration_secs': 0.353901} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.976835] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 733.983766] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.984245] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.984364] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 733.984580] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c626e0b-b9e0-42a4-8136-8dfb3e8e061c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.989272] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Waiting for the task: (returnval){ [ 733.989272] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5287a1d3-6a18-241b-52e3-1639fe7e3fd1" [ 733.989272] env[62569]: _type = "Task" [ 733.989272] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.996729] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5287a1d3-6a18-241b-52e3-1639fe7e3fd1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.083333] env[62569]: DEBUG nova.compute.utils [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 734.086672] env[62569]: DEBUG nova.compute.manager [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 734.086835] env[62569]: DEBUG nova.network.neutron [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 734.179946] env[62569]: DEBUG nova.policy [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5695bbf5e890488a93376ea6994158c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4ec74da9e2c0405cbf85e70f2a8c18e2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 734.342986] env[62569]: INFO nova.scheduler.client.report [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Deleted allocations for instance 311e46eb-85ca-4262-a0a4-21eb073e0dc4 [ 734.437492] env[62569]: DEBUG nova.network.neutron [req-d4497ce9-6bce-4cc5-972a-b0a56c99215e req-e536d47e-bf06-4dac-900f-65dd373f835f service nova] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Updated VIF entry in instance network info cache for port eb2d33a3-7bed-46d7-9656-90aa7577e5e2. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 734.437848] env[62569]: DEBUG nova.network.neutron [req-d4497ce9-6bce-4cc5-972a-b0a56c99215e req-e536d47e-bf06-4dac-900f-65dd373f835f service nova] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Updating instance_info_cache with network_info: [{"id": "eb2d33a3-7bed-46d7-9656-90aa7577e5e2", "address": "fa:16:3e:9e:8d:2c", "network": {"id": "6f2a7189-6586-4180-86d5-ec1324a82e25", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-881682408-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "074644d447544fb794da91d2ce2b0c6d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61a172ee-af3f-473e-b12a-3fee5bf39c8d", "external-id": "nsx-vlan-transportzone-997", "segmentation_id": 997, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb2d33a3-7b", "ovs_interfaceid": "eb2d33a3-7bed-46d7-9656-90aa7577e5e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.499125] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5287a1d3-6a18-241b-52e3-1639fe7e3fd1, 'name': SearchDatastore_Task, 'duration_secs': 0.011378} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.499427] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.499661] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 734.499896] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.500720] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.500720] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 734.500720] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-85fa9b6c-2ddb-41de-a30e-e6cad2d9d5ae {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.508052] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 734.508260] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 734.508954] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-630f5ac8-b3a5-4a5d-86ad-875b9797fb6b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.513967] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Waiting for the task: (returnval){ [ 734.513967] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]526b86c9-efda-dc29-5e55-fe65c9342882" [ 734.513967] env[62569]: _type = "Task" [ 734.513967] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.521102] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]526b86c9-efda-dc29-5e55-fe65c9342882, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.589056] env[62569]: DEBUG nova.compute.manager [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 734.600850] env[62569]: DEBUG nova.network.neutron [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Successfully created port: b0b6ffe7-e392-40a8-b423-566a269133c2 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 734.853125] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cf5f2a5a-79f6-4bdd-b826-3af02dc7c042 tempest-ServersTestJSON-1360786528 tempest-ServersTestJSON-1360786528-project-member] Lock "311e46eb-85ca-4262-a0a4-21eb073e0dc4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.371s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.930544] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd06e799-378e-48dc-acec-dedf30c0ff93 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.938166] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84aef982-ec58-4171-818b-330fb992e487 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.941939] env[62569]: DEBUG oslo_concurrency.lockutils [req-d4497ce9-6bce-4cc5-972a-b0a56c99215e req-e536d47e-bf06-4dac-900f-65dd373f835f service nova] Releasing lock "refresh_cache-93c8445d-689c-4f77-836f-95eafafbb6a2" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.969110] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6fda61-6587-48da-b9b9-c1f8163fa2c6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.976469] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acc01e04-26ce-42a2-bcc7-641c3db7e9a7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.989890] env[62569]: DEBUG nova.compute.provider_tree [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.023886] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]526b86c9-efda-dc29-5e55-fe65c9342882, 'name': SearchDatastore_Task, 'duration_secs': 0.008704} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.024843] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b3f5543-9aed-4209-bbd2-8ae7fa945642 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.029485] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Waiting for the task: (returnval){ [ 735.029485] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]525072bb-76ed-fbff-8d33-f8003760fb58" [ 735.029485] env[62569]: _type = "Task" [ 735.029485] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.037215] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]525072bb-76ed-fbff-8d33-f8003760fb58, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.355779] env[62569]: DEBUG nova.compute.manager [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 735.492698] env[62569]: DEBUG nova.scheduler.client.report [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 735.546200] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]525072bb-76ed-fbff-8d33-f8003760fb58, 'name': SearchDatastore_Task, 'duration_secs': 0.008575} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.546528] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.546736] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 93c8445d-689c-4f77-836f-95eafafbb6a2/93c8445d-689c-4f77-836f-95eafafbb6a2.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 735.547049] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-131b3cf2-8524-4ad2-bad5-e4bc5b969b0b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.554899] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Waiting for the task: (returnval){ [ 735.554899] env[62569]: value = "task-1249871" [ 735.554899] env[62569]: _type = "Task" [ 735.554899] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.563382] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Task: {'id': task-1249871, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.604462] env[62569]: DEBUG nova.compute.manager [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 735.638818] env[62569]: DEBUG nova.virt.hardware [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 735.639078] env[62569]: DEBUG nova.virt.hardware [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 735.639248] env[62569]: DEBUG nova.virt.hardware [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 735.639435] env[62569]: DEBUG nova.virt.hardware [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 735.639558] env[62569]: DEBUG nova.virt.hardware [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 735.639704] env[62569]: DEBUG nova.virt.hardware [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 735.639908] env[62569]: DEBUG nova.virt.hardware [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 735.640126] env[62569]: DEBUG nova.virt.hardware [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 735.640344] env[62569]: DEBUG nova.virt.hardware [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 735.640514] env[62569]: DEBUG nova.virt.hardware [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 735.640683] env[62569]: DEBUG nova.virt.hardware [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 735.641621] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8447cf97-7676-4195-a18a-753985e95b75 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.652065] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6201ba4-e6f9-4aaf-a380-9650a7a57a1f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.894674] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.997875] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.421s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.998523] env[62569]: DEBUG nova.compute.manager [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 736.001610] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.019s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.068029] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Task: {'id': task-1249871, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.453385} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.068029] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 93c8445d-689c-4f77-836f-95eafafbb6a2/93c8445d-689c-4f77-836f-95eafafbb6a2.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 736.068029] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 736.068029] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-75909db6-0567-4cac-ae24-23bce2079722 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.075017] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Waiting for the task: (returnval){ [ 736.075017] env[62569]: value = "task-1249872" [ 736.075017] env[62569]: _type = "Task" [ 736.075017] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.087786] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Task: {'id': task-1249872, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.097860] env[62569]: DEBUG nova.compute.manager [req-1eacd48d-9d3e-44dd-abcc-93491bf726a6 req-5304ab34-6111-4e78-9bb9-64c8a3b8dcbe service nova] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Received event network-vif-plugged-b0b6ffe7-e392-40a8-b423-566a269133c2 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 736.098152] env[62569]: DEBUG oslo_concurrency.lockutils [req-1eacd48d-9d3e-44dd-abcc-93491bf726a6 req-5304ab34-6111-4e78-9bb9-64c8a3b8dcbe service nova] Acquiring lock "3f504d5c-3431-4a80-a7d9-e52af01cf24e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.098403] env[62569]: DEBUG oslo_concurrency.lockutils [req-1eacd48d-9d3e-44dd-abcc-93491bf726a6 req-5304ab34-6111-4e78-9bb9-64c8a3b8dcbe service nova] Lock "3f504d5c-3431-4a80-a7d9-e52af01cf24e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.098612] env[62569]: DEBUG oslo_concurrency.lockutils [req-1eacd48d-9d3e-44dd-abcc-93491bf726a6 req-5304ab34-6111-4e78-9bb9-64c8a3b8dcbe service nova] Lock "3f504d5c-3431-4a80-a7d9-e52af01cf24e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.098820] env[62569]: DEBUG nova.compute.manager [req-1eacd48d-9d3e-44dd-abcc-93491bf726a6 req-5304ab34-6111-4e78-9bb9-64c8a3b8dcbe service nova] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] No waiting events found dispatching network-vif-plugged-b0b6ffe7-e392-40a8-b423-566a269133c2 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 736.099029] env[62569]: WARNING nova.compute.manager [req-1eacd48d-9d3e-44dd-abcc-93491bf726a6 req-5304ab34-6111-4e78-9bb9-64c8a3b8dcbe service nova] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Received unexpected event network-vif-plugged-b0b6ffe7-e392-40a8-b423-566a269133c2 for instance with vm_state building and task_state spawning. [ 736.155228] env[62569]: DEBUG nova.network.neutron [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Successfully updated port: b0b6ffe7-e392-40a8-b423-566a269133c2 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 736.507031] env[62569]: DEBUG nova.compute.utils [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 736.511434] env[62569]: DEBUG nova.compute.manager [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 736.511434] env[62569]: DEBUG nova.network.neutron [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 736.555776] env[62569]: DEBUG nova.policy [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '75adf33d63ac4d07b358ed72daf8f12d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e9ea55b00b574a7db0530efcb498ca2b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 736.582942] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Task: {'id': task-1249872, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.05992} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.583244] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 736.583993] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43df7880-06a6-417b-812a-d369e21555d9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.611865] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] 93c8445d-689c-4f77-836f-95eafafbb6a2/93c8445d-689c-4f77-836f-95eafafbb6a2.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 736.614688] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-881f299c-08ae-4296-b4e2-0f09d1101a5c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.638142] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Waiting for the task: (returnval){ [ 736.638142] env[62569]: value = "task-1249873" [ 736.638142] env[62569]: _type = "Task" [ 736.638142] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.645630] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Task: {'id': task-1249873, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.667580] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Acquiring lock "refresh_cache-3f504d5c-3431-4a80-a7d9-e52af01cf24e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.667580] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Acquired lock "refresh_cache-3f504d5c-3431-4a80-a7d9-e52af01cf24e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.667580] env[62569]: DEBUG nova.network.neutron [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 736.884780] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-173e9200-5ca4-4283-92a8-ba76e323d300 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.893205] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-638acd44-748d-4839-a316-d0c79fcf237a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.933461] env[62569]: DEBUG nova.network.neutron [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Successfully created port: 3da67b46-e89d-4a96-8dd9-752956b9b270 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 736.935993] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92e35352-6224-41bb-a9ff-f5efdc2cac93 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.944146] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55af5d42-5d6f-44ff-9a3b-a01d689709d5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.958437] env[62569]: DEBUG nova.compute.provider_tree [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.012747] env[62569]: DEBUG nova.compute.manager [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 737.148706] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Task: {'id': task-1249873, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.234630] env[62569]: DEBUG nova.network.neutron [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.464288] env[62569]: DEBUG nova.scheduler.client.report [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 737.546476] env[62569]: DEBUG nova.network.neutron [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Updating instance_info_cache with network_info: [{"id": "b0b6ffe7-e392-40a8-b423-566a269133c2", "address": "fa:16:3e:be:fd:ea", "network": {"id": "c05a496b-7346-43b1-983f-dcc11908faa4", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.39", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "3c6a185eb8064b99b7009ae3b3fe4b5a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f408ce42-3cac-4d9d-9c05-15471d653a18", "external-id": "nsx-vlan-transportzone-265", "segmentation_id": 265, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0b6ffe7-e3", "ovs_interfaceid": "b0b6ffe7-e392-40a8-b423-566a269133c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.647468] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Task: {'id': task-1249873, 'name': ReconfigVM_Task, 'duration_secs': 0.771527} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.647760] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Reconfigured VM instance instance-00000032 to attach disk [datastore1] 93c8445d-689c-4f77-836f-95eafafbb6a2/93c8445d-689c-4f77-836f-95eafafbb6a2.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 737.648393] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a045a2bb-c3d5-4544-820a-3818be06d4a9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.655271] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Waiting for the task: (returnval){ [ 737.655271] env[62569]: value = "task-1249875" [ 737.655271] env[62569]: _type = "Task" [ 737.655271] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.666730] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Task: {'id': task-1249875, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.971612] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.970s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.972278] env[62569]: ERROR nova.compute.manager [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5e1b59e6-be6d-4e60-a176-59fe81a57b29, please check neutron logs for more information. [ 737.972278] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Traceback (most recent call last): [ 737.972278] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 737.972278] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] self.driver.spawn(context, instance, image_meta, [ 737.972278] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 737.972278] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 737.972278] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 737.972278] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] vm_ref = self.build_virtual_machine(instance, [ 737.972278] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 737.972278] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] vif_infos = vmwarevif.get_vif_info(self._session, [ 737.972278] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 737.972661] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] for vif in network_info: [ 737.972661] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 737.972661] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] return self._sync_wrapper(fn, *args, **kwargs) [ 737.972661] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 737.972661] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] self.wait() [ 737.972661] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 737.972661] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] self[:] = self._gt.wait() [ 737.972661] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 737.972661] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] return self._exit_event.wait() [ 737.972661] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 737.972661] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] current.throw(*self._exc) [ 737.972661] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.972661] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] result = function(*args, **kwargs) [ 737.973091] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 737.973091] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] return func(*args, **kwargs) [ 737.973091] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 737.973091] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] raise e [ 737.973091] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 737.973091] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] nwinfo = self.network_api.allocate_for_instance( [ 737.973091] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.973091] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] created_port_ids = self._update_ports_for_instance( [ 737.973091] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.973091] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] with excutils.save_and_reraise_exception(): [ 737.973091] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.973091] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] self.force_reraise() [ 737.973091] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.973510] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] raise self.value [ 737.973510] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.973510] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] updated_port = self._update_port( [ 737.973510] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.973510] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] _ensure_no_port_binding_failure(port) [ 737.973510] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.973510] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] raise exception.PortBindingFailed(port_id=port['id']) [ 737.973510] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] nova.exception.PortBindingFailed: Binding failed for port 5e1b59e6-be6d-4e60-a176-59fe81a57b29, please check neutron logs for more information. [ 737.973510] env[62569]: ERROR nova.compute.manager [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] [ 737.973510] env[62569]: DEBUG nova.compute.utils [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Binding failed for port 5e1b59e6-be6d-4e60-a176-59fe81a57b29, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 737.975204] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 15.220s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.975398] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.975639] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62569) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 737.975949] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.187s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.977426] env[62569]: INFO nova.compute.claims [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 737.981094] env[62569]: DEBUG nova.compute.manager [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Build of instance 17f0965b-edf8-4590-9402-2654d5d19f4e was re-scheduled: Binding failed for port 5e1b59e6-be6d-4e60-a176-59fe81a57b29, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 737.982279] env[62569]: DEBUG nova.compute.manager [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 737.982279] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Acquiring lock "refresh_cache-17f0965b-edf8-4590-9402-2654d5d19f4e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.982279] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Acquired lock "refresh_cache-17f0965b-edf8-4590-9402-2654d5d19f4e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.982279] env[62569]: DEBUG nova.network.neutron [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 737.983983] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8608825e-ebba-40d4-bc32-873fc222c9da {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.992460] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96499ac5-694a-4e4f-9b1b-7d8d7ca8d1be {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.006827] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cb884e3-505f-491e-b9f2-ef6279f9cfc1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.013343] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bf4448d-835e-4b46-86dd-eb042b6477e3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.043659] env[62569]: DEBUG nova.compute.manager [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 738.047045] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181543MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=62569) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 738.047193] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.049133] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Releasing lock "refresh_cache-3f504d5c-3431-4a80-a7d9-e52af01cf24e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.049408] env[62569]: DEBUG nova.compute.manager [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Instance network_info: |[{"id": "b0b6ffe7-e392-40a8-b423-566a269133c2", "address": "fa:16:3e:be:fd:ea", "network": {"id": "c05a496b-7346-43b1-983f-dcc11908faa4", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.39", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "3c6a185eb8064b99b7009ae3b3fe4b5a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f408ce42-3cac-4d9d-9c05-15471d653a18", "external-id": "nsx-vlan-transportzone-265", "segmentation_id": 265, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0b6ffe7-e3", "ovs_interfaceid": "b0b6ffe7-e392-40a8-b423-566a269133c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 738.049772] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:be:fd:ea', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f408ce42-3cac-4d9d-9c05-15471d653a18', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b0b6ffe7-e392-40a8-b423-566a269133c2', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 738.057679] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Creating folder: Project (4ec74da9e2c0405cbf85e70f2a8c18e2). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 738.058186] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f2bb1bc5-46f7-47be-bd22-98b174b8fff4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.067648] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Created folder: Project (4ec74da9e2c0405cbf85e70f2a8c18e2) in parent group-v269330. [ 738.067832] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Creating folder: Instances. Parent ref: group-v269353. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 738.069844] env[62569]: DEBUG nova.virt.hardware [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 738.070079] env[62569]: DEBUG nova.virt.hardware [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 738.070306] env[62569]: DEBUG nova.virt.hardware [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 738.070499] env[62569]: DEBUG nova.virt.hardware [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 738.070643] env[62569]: DEBUG nova.virt.hardware [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 738.070787] env[62569]: DEBUG nova.virt.hardware [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 738.070994] env[62569]: DEBUG nova.virt.hardware [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 738.071268] env[62569]: DEBUG nova.virt.hardware [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 738.071451] env[62569]: DEBUG nova.virt.hardware [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 738.071617] env[62569]: DEBUG nova.virt.hardware [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 738.071788] env[62569]: DEBUG nova.virt.hardware [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 738.072033] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3e9a4091-1c3e-47d3-b246-97fd3113ca97 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.074236] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eab9ebd-3f64-4d7d-abee-2f6591f7d0cc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.083124] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b4a1ea-633d-4450-b552-b1ccc3f353ea {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.089163] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Created folder: Instances in parent group-v269353. [ 738.089545] env[62569]: DEBUG oslo.service.loopingcall [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 738.090188] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 738.090762] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-41ad5afb-7211-416e-8c12-1c6d80be2d1c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.119036] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 738.119036] env[62569]: value = "task-1249878" [ 738.119036] env[62569]: _type = "Task" [ 738.119036] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.124595] env[62569]: DEBUG nova.compute.manager [req-fc0acde0-9ebb-4989-a867-11c809027181 req-8de30e1c-63d0-47d6-a0e1-c362aa8c1d32 service nova] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Received event network-changed-b0b6ffe7-e392-40a8-b423-566a269133c2 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 738.124752] env[62569]: DEBUG nova.compute.manager [req-fc0acde0-9ebb-4989-a867-11c809027181 req-8de30e1c-63d0-47d6-a0e1-c362aa8c1d32 service nova] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Refreshing instance network info cache due to event network-changed-b0b6ffe7-e392-40a8-b423-566a269133c2. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 738.124966] env[62569]: DEBUG oslo_concurrency.lockutils [req-fc0acde0-9ebb-4989-a867-11c809027181 req-8de30e1c-63d0-47d6-a0e1-c362aa8c1d32 service nova] Acquiring lock "refresh_cache-3f504d5c-3431-4a80-a7d9-e52af01cf24e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.125123] env[62569]: DEBUG oslo_concurrency.lockutils [req-fc0acde0-9ebb-4989-a867-11c809027181 req-8de30e1c-63d0-47d6-a0e1-c362aa8c1d32 service nova] Acquired lock "refresh_cache-3f504d5c-3431-4a80-a7d9-e52af01cf24e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.125339] env[62569]: DEBUG nova.network.neutron [req-fc0acde0-9ebb-4989-a867-11c809027181 req-8de30e1c-63d0-47d6-a0e1-c362aa8c1d32 service nova] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Refreshing network info cache for port b0b6ffe7-e392-40a8-b423-566a269133c2 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 738.132766] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249878, 'name': CreateVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.164782] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Task: {'id': task-1249875, 'name': Rename_Task, 'duration_secs': 0.145281} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.165080] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 738.165352] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-50e7affd-9f6c-43b7-81f9-73ffa907052a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.171553] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Waiting for the task: (returnval){ [ 738.171553] env[62569]: value = "task-1249879" [ 738.171553] env[62569]: _type = "Task" [ 738.171553] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.179336] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Task: {'id': task-1249879, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.504178] env[62569]: DEBUG nova.network.neutron [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.629398] env[62569]: DEBUG nova.network.neutron [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.635886] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249878, 'name': CreateVM_Task, 'duration_secs': 0.371486} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.636391] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 738.638973] env[62569]: DEBUG oslo_vmware.service [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-365ee462-34bf-429a-af5d-d27114eec4dc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.644579] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.644749] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.645166] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 738.645657] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a338fbf-976d-4691-aaeb-6230f2d4f7b6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.650991] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Waiting for the task: (returnval){ [ 738.650991] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b17302-9ced-9ef0-62ff-0c036d6380fd" [ 738.650991] env[62569]: _type = "Task" [ 738.650991] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.658624] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b17302-9ced-9ef0-62ff-0c036d6380fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.688756] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Task: {'id': task-1249879, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.008553] env[62569]: DEBUG nova.network.neutron [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Successfully updated port: 3da67b46-e89d-4a96-8dd9-752956b9b270 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 739.023592] env[62569]: DEBUG nova.network.neutron [req-fc0acde0-9ebb-4989-a867-11c809027181 req-8de30e1c-63d0-47d6-a0e1-c362aa8c1d32 service nova] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Updated VIF entry in instance network info cache for port b0b6ffe7-e392-40a8-b423-566a269133c2. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 739.023926] env[62569]: DEBUG nova.network.neutron [req-fc0acde0-9ebb-4989-a867-11c809027181 req-8de30e1c-63d0-47d6-a0e1-c362aa8c1d32 service nova] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Updating instance_info_cache with network_info: [{"id": "b0b6ffe7-e392-40a8-b423-566a269133c2", "address": "fa:16:3e:be:fd:ea", "network": {"id": "c05a496b-7346-43b1-983f-dcc11908faa4", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.39", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "3c6a185eb8064b99b7009ae3b3fe4b5a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f408ce42-3cac-4d9d-9c05-15471d653a18", "external-id": "nsx-vlan-transportzone-265", "segmentation_id": 265, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0b6ffe7-e3", "ovs_interfaceid": "b0b6ffe7-e392-40a8-b423-566a269133c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.133787] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Releasing lock "refresh_cache-17f0965b-edf8-4590-9402-2654d5d19f4e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.134043] env[62569]: DEBUG nova.compute.manager [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 739.134201] env[62569]: DEBUG nova.compute.manager [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 739.134374] env[62569]: DEBUG nova.network.neutron [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 739.152137] env[62569]: DEBUG nova.network.neutron [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.167969] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.168255] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 739.168497] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.168638] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.168810] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 739.169533] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1488c35d-7c84-45be-a350-e3ab14edb7e1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.181436] env[62569]: DEBUG oslo_vmware.api [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Task: {'id': task-1249879, 'name': PowerOnVM_Task, 'duration_secs': 0.747597} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.184439] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 739.184648] env[62569]: INFO nova.compute.manager [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Took 7.94 seconds to spawn the instance on the hypervisor. [ 739.184827] env[62569]: DEBUG nova.compute.manager [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 739.185125] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 739.185287] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 739.186458] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c852482-b510-4d32-b06f-5421d0427664 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.189224] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c8f89d4-0839-4e61-ad63-89586e76f5c2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.201087] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b798589f-f334-431d-9b31-607df75e4e9d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.206048] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Waiting for the task: (returnval){ [ 739.206048] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5202077c-3025-8640-c906-ac0bb657dd7c" [ 739.206048] env[62569]: _type = "Task" [ 739.206048] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.215148] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5202077c-3025-8640-c906-ac0bb657dd7c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.316940] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6303631e-18c9-4805-8c5b-735fda9da0f5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.324524] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c0895b-59c7-4886-8cc1-748fc1b362ca {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.355686] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e703ab36-1956-4b9b-a596-8fff22362703 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.363329] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51260eb0-8a8e-482b-ad0e-a819c11618fa {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.376423] env[62569]: DEBUG nova.compute.provider_tree [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.512114] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "refresh_cache-a0b406c3-9466-41bd-9de1-e675cab2ceef" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.512228] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquired lock "refresh_cache-a0b406c3-9466-41bd-9de1-e675cab2ceef" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.512365] env[62569]: DEBUG nova.network.neutron [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 739.526866] env[62569]: DEBUG oslo_concurrency.lockutils [req-fc0acde0-9ebb-4989-a867-11c809027181 req-8de30e1c-63d0-47d6-a0e1-c362aa8c1d32 service nova] Releasing lock "refresh_cache-3f504d5c-3431-4a80-a7d9-e52af01cf24e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.657845] env[62569]: DEBUG nova.network.neutron [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.712675] env[62569]: INFO nova.compute.manager [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Took 27.48 seconds to build instance. [ 739.723889] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Preparing fetch location {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 739.723889] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Creating directory with path [datastore2] vmware_temp/32e28172-71a0-405f-adb6-244fd1af7c04/f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 739.723889] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1b9e327e-9c01-43b8-82d7-ec1db0a50986 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.758507] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Created directory with path [datastore2] vmware_temp/32e28172-71a0-405f-adb6-244fd1af7c04/f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 739.758730] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Fetch image to [datastore2] vmware_temp/32e28172-71a0-405f-adb6-244fd1af7c04/f77800cf-af0d-4e9c-b312-2c59488f2c7c/tmp-sparse.vmdk {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 739.758904] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Downloading image file data f77800cf-af0d-4e9c-b312-2c59488f2c7c to [datastore2] vmware_temp/32e28172-71a0-405f-adb6-244fd1af7c04/f77800cf-af0d-4e9c-b312-2c59488f2c7c/tmp-sparse.vmdk on the data store datastore2 {{(pid=62569) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 739.759734] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dc9701d-d462-43d2-8afe-e84feeb92789 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.767397] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0200111f-4ce3-464d-be18-6ee76920aa8f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.777265] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72239832-a29a-408f-8c0a-0baee23e0d7e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.810944] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c785a88a-f5b9-4561-8e66-def264f48299 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.817643] env[62569]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-a887adc3-ed33-4a6b-8eec-67229cd68bfe {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.841416] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Downloading image file data f77800cf-af0d-4e9c-b312-2c59488f2c7c to the data store datastore2 {{(pid=62569) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 739.880012] env[62569]: DEBUG nova.scheduler.client.report [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 739.899890] env[62569]: DEBUG oslo_vmware.rw_handles [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/32e28172-71a0-405f-adb6-244fd1af7c04/f77800cf-af0d-4e9c-b312-2c59488f2c7c/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62569) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 740.050814] env[62569]: DEBUG nova.network.neutron [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.159691] env[62569]: DEBUG nova.compute.manager [req-d800be78-6e65-41d6-9038-a0e7a4a34d87 req-0164b5ee-c17b-48a6-b2a8-662499b9b6e5 service nova] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Received event network-vif-plugged-3da67b46-e89d-4a96-8dd9-752956b9b270 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 740.160194] env[62569]: DEBUG oslo_concurrency.lockutils [req-d800be78-6e65-41d6-9038-a0e7a4a34d87 req-0164b5ee-c17b-48a6-b2a8-662499b9b6e5 service nova] Acquiring lock "a0b406c3-9466-41bd-9de1-e675cab2ceef-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.160689] env[62569]: DEBUG oslo_concurrency.lockutils [req-d800be78-6e65-41d6-9038-a0e7a4a34d87 req-0164b5ee-c17b-48a6-b2a8-662499b9b6e5 service nova] Lock "a0b406c3-9466-41bd-9de1-e675cab2ceef-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.161031] env[62569]: DEBUG oslo_concurrency.lockutils [req-d800be78-6e65-41d6-9038-a0e7a4a34d87 req-0164b5ee-c17b-48a6-b2a8-662499b9b6e5 service nova] Lock "a0b406c3-9466-41bd-9de1-e675cab2ceef-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.161500] env[62569]: DEBUG nova.compute.manager [req-d800be78-6e65-41d6-9038-a0e7a4a34d87 req-0164b5ee-c17b-48a6-b2a8-662499b9b6e5 service nova] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] No waiting events found dispatching network-vif-plugged-3da67b46-e89d-4a96-8dd9-752956b9b270 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 740.161800] env[62569]: WARNING nova.compute.manager [req-d800be78-6e65-41d6-9038-a0e7a4a34d87 req-0164b5ee-c17b-48a6-b2a8-662499b9b6e5 service nova] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Received unexpected event network-vif-plugged-3da67b46-e89d-4a96-8dd9-752956b9b270 for instance with vm_state building and task_state spawning. [ 740.162622] env[62569]: DEBUG nova.compute.manager [req-d800be78-6e65-41d6-9038-a0e7a4a34d87 req-0164b5ee-c17b-48a6-b2a8-662499b9b6e5 service nova] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Received event network-changed-3da67b46-e89d-4a96-8dd9-752956b9b270 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 740.162622] env[62569]: DEBUG nova.compute.manager [req-d800be78-6e65-41d6-9038-a0e7a4a34d87 req-0164b5ee-c17b-48a6-b2a8-662499b9b6e5 service nova] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Refreshing instance network info cache due to event network-changed-3da67b46-e89d-4a96-8dd9-752956b9b270. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 740.162622] env[62569]: DEBUG oslo_concurrency.lockutils [req-d800be78-6e65-41d6-9038-a0e7a4a34d87 req-0164b5ee-c17b-48a6-b2a8-662499b9b6e5 service nova] Acquiring lock "refresh_cache-a0b406c3-9466-41bd-9de1-e675cab2ceef" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.164301] env[62569]: INFO nova.compute.manager [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] [instance: 17f0965b-edf8-4590-9402-2654d5d19f4e] Took 1.03 seconds to deallocate network for instance. [ 740.171553] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Acquiring lock "93c8445d-689c-4f77-836f-95eafafbb6a2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.215715] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2021cc1b-3902-4f5d-8316-f1f636cbcb51 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Lock "93c8445d-689c-4f77-836f-95eafafbb6a2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.593s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.216966] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Lock "93c8445d-689c-4f77-836f-95eafafbb6a2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.045s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.217204] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Acquiring lock "93c8445d-689c-4f77-836f-95eafafbb6a2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.217405] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Lock "93c8445d-689c-4f77-836f-95eafafbb6a2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.217564] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Lock "93c8445d-689c-4f77-836f-95eafafbb6a2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.219935] env[62569]: INFO nova.compute.manager [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Terminating instance [ 740.223439] env[62569]: DEBUG nova.network.neutron [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Updating instance_info_cache with network_info: [{"id": "3da67b46-e89d-4a96-8dd9-752956b9b270", "address": "fa:16:3e:f6:2c:c2", "network": {"id": "ba6b71b1-bd7b-493e-8682-68d842fa935f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1892479706-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e9ea55b00b574a7db0530efcb498ca2b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3da67b46-e8", "ovs_interfaceid": "3da67b46-e89d-4a96-8dd9-752956b9b270", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.387707] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.412s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.388261] env[62569]: DEBUG nova.compute.manager [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 740.390788] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.731s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.392204] env[62569]: INFO nova.compute.claims [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 740.564606] env[62569]: DEBUG oslo_vmware.rw_handles [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Completed reading data from the image iterator. {{(pid=62569) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 740.564852] env[62569]: DEBUG oslo_vmware.rw_handles [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Closing write handle for https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/32e28172-71a0-405f-adb6-244fd1af7c04/f77800cf-af0d-4e9c-b312-2c59488f2c7c/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62569) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 740.714621] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Downloaded image file data f77800cf-af0d-4e9c-b312-2c59488f2c7c to vmware_temp/32e28172-71a0-405f-adb6-244fd1af7c04/f77800cf-af0d-4e9c-b312-2c59488f2c7c/tmp-sparse.vmdk on the data store datastore2 {{(pid=62569) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 740.717027] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Caching image {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 740.717284] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Copying Virtual Disk [datastore2] vmware_temp/32e28172-71a0-405f-adb6-244fd1af7c04/f77800cf-af0d-4e9c-b312-2c59488f2c7c/tmp-sparse.vmdk to [datastore2] vmware_temp/32e28172-71a0-405f-adb6-244fd1af7c04/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 740.717546] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-26350078-93fd-4536-b86d-26250ced5151 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.721978] env[62569]: DEBUG nova.compute.manager [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 740.725407] env[62569]: DEBUG nova.compute.manager [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 740.725921] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 740.727358] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Releasing lock "refresh_cache-a0b406c3-9466-41bd-9de1-e675cab2ceef" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.727635] env[62569]: DEBUG nova.compute.manager [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Instance network_info: |[{"id": "3da67b46-e89d-4a96-8dd9-752956b9b270", "address": "fa:16:3e:f6:2c:c2", "network": {"id": "ba6b71b1-bd7b-493e-8682-68d842fa935f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1892479706-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e9ea55b00b574a7db0530efcb498ca2b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3da67b46-e8", "ovs_interfaceid": "3da67b46-e89d-4a96-8dd9-752956b9b270", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 740.728950] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50024a43-2d1f-4597-9922-7f7ac8285673 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.731960] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Waiting for the task: (returnval){ [ 740.731960] env[62569]: value = "task-1249881" [ 740.731960] env[62569]: _type = "Task" [ 740.731960] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.731960] env[62569]: DEBUG oslo_concurrency.lockutils [req-d800be78-6e65-41d6-9038-a0e7a4a34d87 req-0164b5ee-c17b-48a6-b2a8-662499b9b6e5 service nova] Acquired lock "refresh_cache-a0b406c3-9466-41bd-9de1-e675cab2ceef" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.731960] env[62569]: DEBUG nova.network.neutron [req-d800be78-6e65-41d6-9038-a0e7a4a34d87 req-0164b5ee-c17b-48a6-b2a8-662499b9b6e5 service nova] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Refreshing network info cache for port 3da67b46-e89d-4a96-8dd9-752956b9b270 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 740.732740] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:2c:c2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8abee039-d93e-48a7-8911-6416a3e1ff30', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3da67b46-e89d-4a96-8dd9-752956b9b270', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 740.740623] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Creating folder: Project (e9ea55b00b574a7db0530efcb498ca2b). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 740.741853] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-598c208d-9d7a-4eec-800e-cbdc17c29c20 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.750829] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 740.752734] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-61c08d1e-cd53-4645-980c-2f87c9f457ee {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.761796] env[62569]: DEBUG oslo_vmware.api [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Waiting for the task: (returnval){ [ 740.761796] env[62569]: value = "task-1249883" [ 740.761796] env[62569]: _type = "Task" [ 740.761796] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.763652] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Created folder: Project (e9ea55b00b574a7db0530efcb498ca2b) in parent group-v269330. [ 740.763652] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Creating folder: Instances. Parent ref: group-v269357. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 740.767396] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4006acb9-badf-4fa9-949f-8d5eb47427d5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.773205] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': task-1249881, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.777688] env[62569]: DEBUG oslo_vmware.api [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Task: {'id': task-1249883, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.783977] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Created folder: Instances in parent group-v269357. [ 740.784194] env[62569]: DEBUG oslo.service.loopingcall [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 740.784396] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 740.784535] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0cb6cd4c-26f5-408f-b478-1c0e482a2ebd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.805741] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 740.805741] env[62569]: value = "task-1249886" [ 740.805741] env[62569]: _type = "Task" [ 740.805741] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.814955] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249886, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.897410] env[62569]: DEBUG nova.compute.utils [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 740.902052] env[62569]: DEBUG nova.compute.manager [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Not allocating networking since 'none' was specified. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 741.195397] env[62569]: INFO nova.scheduler.client.report [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Deleted allocations for instance 17f0965b-edf8-4590-9402-2654d5d19f4e [ 741.246902] env[62569]: DEBUG oslo_concurrency.lockutils [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.260925] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': task-1249881, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.275631] env[62569]: DEBUG oslo_vmware.api [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Task: {'id': task-1249883, 'name': PowerOffVM_Task, 'duration_secs': 0.223221} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.275929] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 741.276131] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 741.276433] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-549b60ed-e27d-4489-96be-8c4f3a24d988 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.315841] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249886, 'name': CreateVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.361924] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 741.362343] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 741.362589] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Deleting the datastore file [datastore1] 93c8445d-689c-4f77-836f-95eafafbb6a2 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 741.363063] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a46cb4d7-f280-4f66-9e83-4aae3fe027b8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.373659] env[62569]: DEBUG oslo_vmware.api [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Waiting for the task: (returnval){ [ 741.373659] env[62569]: value = "task-1249888" [ 741.373659] env[62569]: _type = "Task" [ 741.373659] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.383938] env[62569]: DEBUG oslo_vmware.api [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Task: {'id': task-1249888, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.402809] env[62569]: DEBUG nova.compute.manager [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 741.485274] env[62569]: DEBUG nova.network.neutron [req-d800be78-6e65-41d6-9038-a0e7a4a34d87 req-0164b5ee-c17b-48a6-b2a8-662499b9b6e5 service nova] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Updated VIF entry in instance network info cache for port 3da67b46-e89d-4a96-8dd9-752956b9b270. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 741.485719] env[62569]: DEBUG nova.network.neutron [req-d800be78-6e65-41d6-9038-a0e7a4a34d87 req-0164b5ee-c17b-48a6-b2a8-662499b9b6e5 service nova] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Updating instance_info_cache with network_info: [{"id": "3da67b46-e89d-4a96-8dd9-752956b9b270", "address": "fa:16:3e:f6:2c:c2", "network": {"id": "ba6b71b1-bd7b-493e-8682-68d842fa935f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1892479706-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e9ea55b00b574a7db0530efcb498ca2b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3da67b46-e8", "ovs_interfaceid": "3da67b46-e89d-4a96-8dd9-752956b9b270", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.690174] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db5d1b74-65a3-4229-a854-f2ced56d55ab {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.698404] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af99b986-31d8-4d8c-a0f5-ce4150d307b6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.703381] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2be54b82-eaed-40b5-9214-4eb816c4d745 tempest-FloatingIPsAssociationTestJSON-171995455 tempest-FloatingIPsAssociationTestJSON-171995455-project-member] Lock "17f0965b-edf8-4590-9402-2654d5d19f4e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.912s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.730650] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe65e0bc-2bc3-448e-a6f3-6611c6d8cce8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.739939] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0ea5698-19c9-4cb1-b173-3c809e4ab8aa {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.757750] env[62569]: DEBUG nova.compute.provider_tree [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 741.761774] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': task-1249881, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.766952} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.762272] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Copied Virtual Disk [datastore2] vmware_temp/32e28172-71a0-405f-adb6-244fd1af7c04/f77800cf-af0d-4e9c-b312-2c59488f2c7c/tmp-sparse.vmdk to [datastore2] vmware_temp/32e28172-71a0-405f-adb6-244fd1af7c04/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 741.762458] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Deleting the datastore file [datastore2] vmware_temp/32e28172-71a0-405f-adb6-244fd1af7c04/f77800cf-af0d-4e9c-b312-2c59488f2c7c/tmp-sparse.vmdk {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 741.762702] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-33228728-59ef-4ad5-b54b-3ca214193a33 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.772112] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Waiting for the task: (returnval){ [ 741.772112] env[62569]: value = "task-1249889" [ 741.772112] env[62569]: _type = "Task" [ 741.772112] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.781011] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': task-1249889, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.815463] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249886, 'name': CreateVM_Task, 'duration_secs': 0.534888} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.815633] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 741.816318] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.816480] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.816829] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 741.817030] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df2ef926-eff4-4091-8480-4563bfc5dd3b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.822265] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 741.822265] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52e139ec-cd80-2206-5726-28cf82120c27" [ 741.822265] env[62569]: _type = "Task" [ 741.822265] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.830800] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52e139ec-cd80-2206-5726-28cf82120c27, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.883376] env[62569]: DEBUG oslo_vmware.api [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Task: {'id': task-1249888, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.228248} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.883727] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 741.883960] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 741.884230] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 741.884461] env[62569]: INFO nova.compute.manager [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Took 1.16 seconds to destroy the instance on the hypervisor. [ 741.884753] env[62569]: DEBUG oslo.service.loopingcall [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 741.885044] env[62569]: DEBUG nova.compute.manager [-] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 741.885151] env[62569]: DEBUG nova.network.neutron [-] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 741.992318] env[62569]: DEBUG oslo_concurrency.lockutils [req-d800be78-6e65-41d6-9038-a0e7a4a34d87 req-0164b5ee-c17b-48a6-b2a8-662499b9b6e5 service nova] Releasing lock "refresh_cache-a0b406c3-9466-41bd-9de1-e675cab2ceef" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.208020] env[62569]: DEBUG nova.compute.manager [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 742.224604] env[62569]: DEBUG nova.compute.manager [req-88cca029-623c-445a-ba96-e0e2c6660f90 req-8ab7d370-e599-4fe2-89eb-c79c7e5b3427 service nova] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Received event network-vif-deleted-eb2d33a3-7bed-46d7-9656-90aa7577e5e2 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 742.224816] env[62569]: INFO nova.compute.manager [req-88cca029-623c-445a-ba96-e0e2c6660f90 req-8ab7d370-e599-4fe2-89eb-c79c7e5b3427 service nova] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Neutron deleted interface eb2d33a3-7bed-46d7-9656-90aa7577e5e2; detaching it from the instance and deleting it from the info cache [ 742.224996] env[62569]: DEBUG nova.network.neutron [req-88cca029-623c-445a-ba96-e0e2c6660f90 req-8ab7d370-e599-4fe2-89eb-c79c7e5b3427 service nova] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.264202] env[62569]: DEBUG nova.scheduler.client.report [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 742.282547] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': task-1249889, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.03286} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.283416] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 742.283640] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Moving file from [datastore2] vmware_temp/32e28172-71a0-405f-adb6-244fd1af7c04/f77800cf-af0d-4e9c-b312-2c59488f2c7c to [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c. {{(pid=62569) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 742.283891] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-64ff5e5a-7d47-48e4-a3e7-4f81c5c7a566 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.293642] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Waiting for the task: (returnval){ [ 742.293642] env[62569]: value = "task-1249890" [ 742.293642] env[62569]: _type = "Task" [ 742.293642] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.302168] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': task-1249890, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.331929] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.332187] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 742.332398] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.414956] env[62569]: DEBUG nova.compute.manager [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 742.439656] env[62569]: DEBUG nova.virt.hardware [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 742.439914] env[62569]: DEBUG nova.virt.hardware [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 742.440077] env[62569]: DEBUG nova.virt.hardware [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 742.440328] env[62569]: DEBUG nova.virt.hardware [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 742.440483] env[62569]: DEBUG nova.virt.hardware [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 742.441071] env[62569]: DEBUG nova.virt.hardware [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 742.441071] env[62569]: DEBUG nova.virt.hardware [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 742.441071] env[62569]: DEBUG nova.virt.hardware [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 742.441329] env[62569]: DEBUG nova.virt.hardware [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 742.441473] env[62569]: DEBUG nova.virt.hardware [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 742.444728] env[62569]: DEBUG nova.virt.hardware [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 742.444728] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34501b16-40bf-4ee1-b0ee-88defb6e4ea6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.451389] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9579b976-c24e-4151-be3a-86cc50d1828f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.468431] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Instance VIF info [] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 742.474184] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Creating folder: Project (de04688cdf374f1bb3b5f3b0fa7d6eb7). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 742.474503] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-45a21c4a-1338-4841-93ea-5d250bc80dff {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.487125] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Created folder: Project (de04688cdf374f1bb3b5f3b0fa7d6eb7) in parent group-v269330. [ 742.487385] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Creating folder: Instances. Parent ref: group-v269360. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 742.487570] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-315e5afa-ad35-418d-833f-5e9378930d5c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.502019] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Created folder: Instances in parent group-v269360. [ 742.502019] env[62569]: DEBUG oslo.service.loopingcall [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 742.502019] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 94392d48-223d-4205-9627-25488a468769] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 742.502019] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-00a0f93c-b406-48a2-a0eb-a26bec12d9e7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.517412] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 742.517412] env[62569]: value = "task-1249893" [ 742.517412] env[62569]: _type = "Task" [ 742.517412] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.525405] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249893, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.644699] env[62569]: DEBUG nova.network.neutron [-] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.732926] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-66788ef7-fa57-489d-a41d-751845c09cb4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.741333] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a04d3e2-51ca-4486-8208-48ccbc017aba {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.754279] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.771196] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.380s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.771997] env[62569]: DEBUG nova.compute.manager [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 742.774805] env[62569]: DEBUG nova.compute.manager [req-88cca029-623c-445a-ba96-e0e2c6660f90 req-8ab7d370-e599-4fe2-89eb-c79c7e5b3427 service nova] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Detach interface failed, port_id=eb2d33a3-7bed-46d7-9656-90aa7577e5e2, reason: Instance 93c8445d-689c-4f77-836f-95eafafbb6a2 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 742.776028] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.191s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.805166] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': task-1249890, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.029205} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.806215] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] File moved {{(pid=62569) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 742.806215] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Cleaning up location [datastore2] vmware_temp/32e28172-71a0-405f-adb6-244fd1af7c04 {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 742.806424] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Deleting the datastore file [datastore2] vmware_temp/32e28172-71a0-405f-adb6-244fd1af7c04 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 742.807345] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-00c13eb1-26f9-46b9-89b3-e1e86fb67b13 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.816192] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Waiting for the task: (returnval){ [ 742.816192] env[62569]: value = "task-1249894" [ 742.816192] env[62569]: _type = "Task" [ 742.816192] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.826686] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': task-1249894, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.034725] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249893, 'name': CreateVM_Task, 'duration_secs': 0.329619} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.035014] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 94392d48-223d-4205-9627-25488a468769] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 743.035575] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.035837] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.036293] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 743.036640] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-235d71c1-8e53-4c91-b542-e1f6c0ea6228 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.042854] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 743.042854] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520e5653-7fd0-8ced-47a5-7e02a1d1946b" [ 743.042854] env[62569]: _type = "Task" [ 743.042854] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.051867] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520e5653-7fd0-8ced-47a5-7e02a1d1946b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.147799] env[62569]: INFO nova.compute.manager [-] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Took 1.26 seconds to deallocate network for instance. [ 743.284887] env[62569]: DEBUG nova.compute.utils [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 743.286755] env[62569]: DEBUG nova.compute.manager [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Not allocating networking since 'none' was specified. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 743.326572] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': task-1249894, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.029093} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.326830] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 743.327677] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23addf4d-d05d-4a52-9d7f-968d489b5ed0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.334015] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Waiting for the task: (returnval){ [ 743.334015] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]522cc47b-502e-ac37-31db-4df23a3cba48" [ 743.334015] env[62569]: _type = "Task" [ 743.334015] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.346743] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]522cc47b-502e-ac37-31db-4df23a3cba48, 'name': SearchDatastore_Task, 'duration_secs': 0.009775} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.346992] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.347297] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 3f504d5c-3431-4a80-a7d9-e52af01cf24e/3f504d5c-3431-4a80-a7d9-e52af01cf24e.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 743.347568] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.347747] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 743.347970] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3d7d3662-d9f8-4e8c-86a5-6da23d59b8bb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.349989] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7b2860c4-8ec9-41c2-8d84-2d19603ce595 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.361695] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Waiting for the task: (returnval){ [ 743.361695] env[62569]: value = "task-1249896" [ 743.361695] env[62569]: _type = "Task" [ 743.361695] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.365980] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 743.366187] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 743.369793] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e7f1a3d-acc4-447b-8902-4bdfe8ce5388 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.375447] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': task-1249896, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.378797] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 743.378797] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f876c3-fca6-d0cb-3f50-1ba42a123ab5" [ 743.378797] env[62569]: _type = "Task" [ 743.378797] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.387713] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f876c3-fca6-d0cb-3f50-1ba42a123ab5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.557310] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520e5653-7fd0-8ced-47a5-7e02a1d1946b, 'name': SearchDatastore_Task, 'duration_secs': 0.011464} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.557310] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.557310] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 743.557310] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.557477] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.558379] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 743.561830] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5da872df-e0ab-48a6-ae51-a2d80b110aa5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.574456] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 743.576060] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 743.576060] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83a70fea-1551-4739-8e6a-5a3c1bb0cbee {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.589029] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 743.589029] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520b9953-591f-224a-4545-a204f8308a3f" [ 743.589029] env[62569]: _type = "Task" [ 743.589029] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.599445] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520b9953-591f-224a-4545-a204f8308a3f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.643098] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69e1525f-c93f-4640-8b4c-25a1166011b7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.656329] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.656329] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db6f4955-a414-444e-b389-b29ec15a88b1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.695566] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d7b866f-13ff-4aaa-9b53-7d15468a6271 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.706807] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43ce1af7-407b-4602-b42c-627f7437f17b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.727865] env[62569]: DEBUG nova.compute.provider_tree [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.788421] env[62569]: DEBUG nova.compute.manager [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 743.871541] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': task-1249896, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45239} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.871912] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 3f504d5c-3431-4a80-a7d9-e52af01cf24e/3f504d5c-3431-4a80-a7d9-e52af01cf24e.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 743.872048] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 743.872302] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2eba486e-9e7d-4d6f-bb48-3fa1bac72550 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.880054] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Waiting for the task: (returnval){ [ 743.880054] env[62569]: value = "task-1249897" [ 743.880054] env[62569]: _type = "Task" [ 743.880054] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.893121] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f876c3-fca6-d0cb-3f50-1ba42a123ab5, 'name': SearchDatastore_Task, 'duration_secs': 0.019318} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.896561] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': task-1249897, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.896779] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c55e505-2d66-4f9a-87d2-782e9ad40f0c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.903913] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 743.903913] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]526ff9ff-efe3-7d29-28de-59739627a97b" [ 743.903913] env[62569]: _type = "Task" [ 743.903913] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.912484] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]526ff9ff-efe3-7d29-28de-59739627a97b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.100158] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520b9953-591f-224a-4545-a204f8308a3f, 'name': SearchDatastore_Task, 'duration_secs': 0.018208} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.101015] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0311f92-c848-463e-bf1b-b63ffee29131 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.107899] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 744.107899] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52e227a0-85ff-bcdb-4be7-70d1dbc6b42b" [ 744.107899] env[62569]: _type = "Task" [ 744.107899] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.116169] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52e227a0-85ff-bcdb-4be7-70d1dbc6b42b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.233912] env[62569]: DEBUG nova.scheduler.client.report [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 744.394668] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': task-1249897, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.14885} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.394997] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 744.395700] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe53049d-ddbc-40a1-b3a0-a4ef900c1abc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.419424] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 3f504d5c-3431-4a80-a7d9-e52af01cf24e/3f504d5c-3431-4a80-a7d9-e52af01cf24e.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 744.422587] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-858de145-3f86-452b-ba4a-bdfcef90f382 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.444381] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]526ff9ff-efe3-7d29-28de-59739627a97b, 'name': SearchDatastore_Task, 'duration_secs': 0.010006} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.444688] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.444959] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] a0b406c3-9466-41bd-9de1-e675cab2ceef/a0b406c3-9466-41bd-9de1-e675cab2ceef.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 744.445271] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Waiting for the task: (returnval){ [ 744.445271] env[62569]: value = "task-1249898" [ 744.445271] env[62569]: _type = "Task" [ 744.445271] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.445453] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ae049819-d404-450c-9edd-e1c1fba6b48f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.457227] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': task-1249898, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.458462] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 744.458462] env[62569]: value = "task-1249899" [ 744.458462] env[62569]: _type = "Task" [ 744.458462] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.467695] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1249899, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.624971] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52e227a0-85ff-bcdb-4be7-70d1dbc6b42b, 'name': SearchDatastore_Task, 'duration_secs': 0.009976} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.625452] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.626322] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 94392d48-223d-4205-9627-25488a468769/94392d48-223d-4205-9627-25488a468769.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 744.626322] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7a07b5b3-3e24-4365-969f-b76ab7f2be45 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.639027] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 744.639027] env[62569]: value = "task-1249900" [ 744.639027] env[62569]: _type = "Task" [ 744.639027] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.648996] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249900, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.738039] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.962s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.738609] env[62569]: ERROR nova.compute.manager [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bdb661f2-6f69-4395-a9cc-5fa2b17513b1, please check neutron logs for more information. [ 744.738609] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Traceback (most recent call last): [ 744.738609] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 744.738609] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] self.driver.spawn(context, instance, image_meta, [ 744.738609] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 744.738609] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 744.738609] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 744.738609] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] vm_ref = self.build_virtual_machine(instance, [ 744.738609] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 744.738609] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] vif_infos = vmwarevif.get_vif_info(self._session, [ 744.738609] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 744.739027] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] for vif in network_info: [ 744.739027] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 744.739027] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] return self._sync_wrapper(fn, *args, **kwargs) [ 744.739027] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 744.739027] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] self.wait() [ 744.739027] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 744.739027] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] self[:] = self._gt.wait() [ 744.739027] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 744.739027] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] return self._exit_event.wait() [ 744.739027] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 744.739027] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] current.throw(*self._exc) [ 744.739027] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 744.739027] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] result = function(*args, **kwargs) [ 744.739445] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 744.739445] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] return func(*args, **kwargs) [ 744.739445] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 744.739445] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] raise e [ 744.739445] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 744.739445] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] nwinfo = self.network_api.allocate_for_instance( [ 744.739445] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 744.739445] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] created_port_ids = self._update_ports_for_instance( [ 744.739445] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 744.739445] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] with excutils.save_and_reraise_exception(): [ 744.739445] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.739445] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] self.force_reraise() [ 744.739445] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.739827] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] raise self.value [ 744.739827] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 744.739827] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] updated_port = self._update_port( [ 744.739827] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.739827] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] _ensure_no_port_binding_failure(port) [ 744.739827] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.739827] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] raise exception.PortBindingFailed(port_id=port['id']) [ 744.739827] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] nova.exception.PortBindingFailed: Binding failed for port bdb661f2-6f69-4395-a9cc-5fa2b17513b1, please check neutron logs for more information. [ 744.739827] env[62569]: ERROR nova.compute.manager [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] [ 744.740590] env[62569]: DEBUG nova.compute.utils [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Binding failed for port bdb661f2-6f69-4395-a9cc-5fa2b17513b1, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 744.743201] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.126s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.747965] env[62569]: DEBUG nova.compute.manager [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Build of instance 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2 was re-scheduled: Binding failed for port bdb661f2-6f69-4395-a9cc-5fa2b17513b1, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 744.748882] env[62569]: DEBUG nova.compute.manager [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 744.749297] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Acquiring lock "refresh_cache-3551a2e8-2f7f-475d-9c5c-2c8e88335bc2" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.749495] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Acquired lock "refresh_cache-3551a2e8-2f7f-475d-9c5c-2c8e88335bc2" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.749662] env[62569]: DEBUG nova.network.neutron [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 744.797799] env[62569]: DEBUG nova.compute.manager [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 744.825972] env[62569]: DEBUG nova.virt.hardware [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 744.826254] env[62569]: DEBUG nova.virt.hardware [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 744.826410] env[62569]: DEBUG nova.virt.hardware [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 744.826597] env[62569]: DEBUG nova.virt.hardware [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 744.826745] env[62569]: DEBUG nova.virt.hardware [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 744.826893] env[62569]: DEBUG nova.virt.hardware [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 744.827117] env[62569]: DEBUG nova.virt.hardware [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 744.827284] env[62569]: DEBUG nova.virt.hardware [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 744.827455] env[62569]: DEBUG nova.virt.hardware [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 744.827616] env[62569]: DEBUG nova.virt.hardware [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 744.827788] env[62569]: DEBUG nova.virt.hardware [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 744.828702] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0267891-6838-4d90-86ca-897d3ae64c92 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.839572] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09376d87-2b42-4169-9a9e-15097ea6a579 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.855247] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Instance VIF info [] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 744.860987] env[62569]: DEBUG oslo.service.loopingcall [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 744.861371] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 744.861611] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-44df3a12-2413-4b76-9631-9fcd83acc69c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.882695] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 744.882695] env[62569]: value = "task-1249901" [ 744.882695] env[62569]: _type = "Task" [ 744.882695] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.892977] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249901, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.960796] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': task-1249898, 'name': ReconfigVM_Task, 'duration_secs': 0.351782} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.966332] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 3f504d5c-3431-4a80-a7d9-e52af01cf24e/3f504d5c-3431-4a80-a7d9-e52af01cf24e.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 744.967251] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3acf0dc7-eee5-48d7-900d-e305b9c60cef {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.977908] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1249899, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.493184} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.979912] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] a0b406c3-9466-41bd-9de1-e675cab2ceef/a0b406c3-9466-41bd-9de1-e675cab2ceef.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 744.981225] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 744.982069] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Waiting for the task: (returnval){ [ 744.982069] env[62569]: value = "task-1249902" [ 744.982069] env[62569]: _type = "Task" [ 744.982069] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.982366] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a7dae5aa-8321-4102-848a-8701ea79471c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.001443] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 745.001443] env[62569]: value = "task-1249904" [ 745.001443] env[62569]: _type = "Task" [ 745.001443] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.001867] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': task-1249902, 'name': Rename_Task} progress is 10%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.014834] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1249904, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.148904] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249900, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.275578] env[62569]: DEBUG nova.network.neutron [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.396617] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249901, 'name': CreateVM_Task} progress is 25%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.450807] env[62569]: DEBUG nova.network.neutron [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.500441] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': task-1249902, 'name': Rename_Task, 'duration_secs': 0.180071} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.501057] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 745.501326] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-203572bc-ba2c-4d9e-b412-1f975f6d899b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.513955] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1249904, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100845} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.515270] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 745.515627] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Waiting for the task: (returnval){ [ 745.515627] env[62569]: value = "task-1249905" [ 745.515627] env[62569]: _type = "Task" [ 745.515627] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.516285] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-794297cd-0830-443e-8ea3-4526b687df2a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.529069] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': task-1249905, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.546900] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] a0b406c3-9466-41bd-9de1-e675cab2ceef/a0b406c3-9466-41bd-9de1-e675cab2ceef.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 745.549641] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac98d1b4-b60e-4cd8-97c1-2c763059d22a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.570853] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 745.570853] env[62569]: value = "task-1249906" [ 745.570853] env[62569]: _type = "Task" [ 745.570853] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.582405] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1249906, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.651319] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249900, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.724542} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.651611] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 94392d48-223d-4205-9627-25488a468769/94392d48-223d-4205-9627-25488a468769.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 745.651837] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 745.652104] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6d87a61d-6b10-4732-a88c-6413c34f89e5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.655183] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-536cae4f-6b48-4b1d-944e-68aea3fb27f1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.662488] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd34527b-4331-4cd6-a67b-49d27cd55bdc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.666485] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 745.666485] env[62569]: value = "task-1249907" [ 745.666485] env[62569]: _type = "Task" [ 745.666485] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.697778] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d19eaa0-310a-412b-ba38-207888784547 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.700060] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249907, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.705728] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a1055b1-2322-433c-b21b-2c8e03452d8f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.721378] env[62569]: DEBUG nova.compute.provider_tree [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 745.896753] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249901, 'name': CreateVM_Task, 'duration_secs': 0.656824} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.896753] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 745.898339] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.898643] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.900824] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 745.900824] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23113ef9-117a-4707-82bb-717a715b6a33 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.905729] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 745.905729] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a3e697-ffeb-2096-306a-5c7305d52f3c" [ 745.905729] env[62569]: _type = "Task" [ 745.905729] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.915282] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a3e697-ffeb-2096-306a-5c7305d52f3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.953181] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Releasing lock "refresh_cache-3551a2e8-2f7f-475d-9c5c-2c8e88335bc2" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.953447] env[62569]: DEBUG nova.compute.manager [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 745.953634] env[62569]: DEBUG nova.compute.manager [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 745.953827] env[62569]: DEBUG nova.network.neutron [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 745.973811] env[62569]: DEBUG nova.network.neutron [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.028833] env[62569]: DEBUG oslo_vmware.api [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': task-1249905, 'name': PowerOnVM_Task, 'duration_secs': 0.510203} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.029275] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 746.029548] env[62569]: INFO nova.compute.manager [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Took 10.42 seconds to spawn the instance on the hypervisor. [ 746.029831] env[62569]: DEBUG nova.compute.manager [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 746.030659] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c7a6147-2d77-4f0b-aa32-1188a7c38fc7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.091120] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1249906, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.177819] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249907, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07916} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.178459] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 746.179515] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-152d818a-9300-4a45-8727-c2cf8cb82626 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.203075] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] 94392d48-223d-4205-9627-25488a468769/94392d48-223d-4205-9627-25488a468769.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 746.203431] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f6da16e-b61c-4cd3-8660-6dc4460b2d19 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.224790] env[62569]: DEBUG nova.scheduler.client.report [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 746.233641] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 746.233641] env[62569]: value = "task-1249908" [ 746.233641] env[62569]: _type = "Task" [ 746.233641] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.242294] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249908, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.420373] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a3e697-ffeb-2096-306a-5c7305d52f3c, 'name': SearchDatastore_Task, 'duration_secs': 0.010418} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.421344] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.421734] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 746.422141] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.422592] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.422807] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 746.423143] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-41804cef-bd92-4e91-ab7d-37e14d1ff28f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.435256] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 746.435500] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 746.436391] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1985f2b2-8ff2-4c89-b3a7-ebc5d0ef71f7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.443182] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 746.443182] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527771de-5b84-0535-9c22-9c7f6c24ca35" [ 746.443182] env[62569]: _type = "Task" [ 746.443182] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.453110] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527771de-5b84-0535-9c22-9c7f6c24ca35, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.475847] env[62569]: DEBUG nova.network.neutron [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.553732] env[62569]: INFO nova.compute.manager [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Took 30.00 seconds to build instance. [ 746.585039] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1249906, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.731412] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.988s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.732152] env[62569]: ERROR nova.compute.manager [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0a71b5f5-004f-44f5-b118-f2f26d548ffa, please check neutron logs for more information. [ 746.732152] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Traceback (most recent call last): [ 746.732152] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 746.732152] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] self.driver.spawn(context, instance, image_meta, [ 746.732152] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 746.732152] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] self._vmops.spawn(context, instance, image_meta, injected_files, [ 746.732152] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 746.732152] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] vm_ref = self.build_virtual_machine(instance, [ 746.732152] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 746.732152] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] vif_infos = vmwarevif.get_vif_info(self._session, [ 746.732152] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 746.732535] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] for vif in network_info: [ 746.732535] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 746.732535] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] return self._sync_wrapper(fn, *args, **kwargs) [ 746.732535] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 746.732535] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] self.wait() [ 746.732535] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 746.732535] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] self[:] = self._gt.wait() [ 746.732535] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 746.732535] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] return self._exit_event.wait() [ 746.732535] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 746.732535] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] current.throw(*self._exc) [ 746.732535] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.732535] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] result = function(*args, **kwargs) [ 746.733083] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 746.733083] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] return func(*args, **kwargs) [ 746.733083] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 746.733083] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] raise e [ 746.733083] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 746.733083] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] nwinfo = self.network_api.allocate_for_instance( [ 746.733083] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.733083] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] created_port_ids = self._update_ports_for_instance( [ 746.733083] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.733083] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] with excutils.save_and_reraise_exception(): [ 746.733083] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.733083] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] self.force_reraise() [ 746.733083] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.733489] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] raise self.value [ 746.733489] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.733489] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] updated_port = self._update_port( [ 746.733489] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.733489] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] _ensure_no_port_binding_failure(port) [ 746.733489] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.733489] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] raise exception.PortBindingFailed(port_id=port['id']) [ 746.733489] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] nova.exception.PortBindingFailed: Binding failed for port 0a71b5f5-004f-44f5-b118-f2f26d548ffa, please check neutron logs for more information. [ 746.733489] env[62569]: ERROR nova.compute.manager [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] [ 746.733489] env[62569]: DEBUG nova.compute.utils [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Binding failed for port 0a71b5f5-004f-44f5-b118-f2f26d548ffa, please check neutron logs for more information. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 746.734234] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.035s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.735719] env[62569]: INFO nova.compute.claims [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 746.739581] env[62569]: DEBUG nova.compute.manager [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Build of instance 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda was re-scheduled: Binding failed for port 0a71b5f5-004f-44f5-b118-f2f26d548ffa, please check neutron logs for more information. {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 746.740022] env[62569]: DEBUG nova.compute.manager [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Unplugging VIFs for instance {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 746.740300] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Acquiring lock "refresh_cache-0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.740449] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Acquired lock "refresh_cache-0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.740611] env[62569]: DEBUG nova.network.neutron [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 746.755439] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249908, 'name': ReconfigVM_Task, 'duration_secs': 0.283727} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.755439] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Reconfigured VM instance instance-00000035 to attach disk [datastore1] 94392d48-223d-4205-9627-25488a468769/94392d48-223d-4205-9627-25488a468769.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 746.755439] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8bab4641-f0a9-40b7-b6ef-4728c28bf7b4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.765192] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 746.765192] env[62569]: value = "task-1249909" [ 746.765192] env[62569]: _type = "Task" [ 746.765192] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.775529] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249909, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.955765] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527771de-5b84-0535-9c22-9c7f6c24ca35, 'name': SearchDatastore_Task, 'duration_secs': 0.01226} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.956794] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7dec23c-fdea-4865-9353-7b7cdcfa0380 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.965679] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 746.965679] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5299c1dd-af0d-3bad-6128-84bbb8d17823" [ 746.965679] env[62569]: _type = "Task" [ 746.965679] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.977468] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5299c1dd-af0d-3bad-6128-84bbb8d17823, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.979381] env[62569]: INFO nova.compute.manager [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] [instance: 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2] Took 1.03 seconds to deallocate network for instance. [ 747.062597] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e47ef252-21ba-4155-81f5-83332caee67d tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Lock "3f504d5c-3431-4a80-a7d9-e52af01cf24e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 138.138s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.086860] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1249906, 'name': ReconfigVM_Task, 'duration_secs': 1.226584} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.089249] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Reconfigured VM instance instance-00000034 to attach disk [datastore2] a0b406c3-9466-41bd-9de1-e675cab2ceef/a0b406c3-9466-41bd-9de1-e675cab2ceef.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 747.089249] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ae46edde-6382-4729-a657-8830029dbc55 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.096962] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 747.096962] env[62569]: value = "task-1249910" [ 747.096962] env[62569]: _type = "Task" [ 747.096962] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.111360] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1249910, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.279391] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249909, 'name': Rename_Task, 'duration_secs': 0.144597} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.279391] env[62569]: DEBUG nova.network.neutron [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.280774] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 747.280881] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-48fd49cd-4bcf-420e-b0a2-b749d7af132e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.289327] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 747.289327] env[62569]: value = "task-1249911" [ 747.289327] env[62569]: _type = "Task" [ 747.289327] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.300140] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249911, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.413026] env[62569]: DEBUG nova.network.neutron [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.477729] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5299c1dd-af0d-3bad-6128-84bbb8d17823, 'name': SearchDatastore_Task, 'duration_secs': 0.010368} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.478063] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.478381] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 015140aa-8336-40e2-856c-723277a48f6e/015140aa-8336-40e2-856c-723277a48f6e.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 747.478690] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-80b8909b-f509-4119-a5e6-a1ac7e2b9dac {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.488646] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 747.488646] env[62569]: value = "task-1249912" [ 747.488646] env[62569]: _type = "Task" [ 747.488646] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.498808] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249912, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.566413] env[62569]: DEBUG nova.compute.manager [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 747.616902] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1249910, 'name': Rename_Task, 'duration_secs': 0.157329} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.617877] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 747.617877] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d3bc3502-6698-452e-9f04-0961c68788af {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.625827] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 747.625827] env[62569]: value = "task-1249913" [ 747.625827] env[62569]: _type = "Task" [ 747.625827] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.467214] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Releasing lock "refresh_cache-0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.467214] env[62569]: DEBUG nova.compute.manager [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62569) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 748.467214] env[62569]: DEBUG nova.compute.manager [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 748.467214] env[62569]: DEBUG nova.network.neutron [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 748.489339] env[62569]: DEBUG oslo_vmware.api [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1249913, 'name': PowerOnVM_Task, 'duration_secs': 0.621851} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.494993] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 748.495644] env[62569]: INFO nova.compute.manager [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Took 10.45 seconds to spawn the instance on the hypervisor. [ 748.495644] env[62569]: DEBUG nova.compute.manager [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 748.495737] env[62569]: DEBUG oslo_vmware.api [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249911, 'name': PowerOnVM_Task, 'duration_secs': 0.501764} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.495873] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249912, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45121} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.497112] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d199b7fd-802d-4575-873a-bb0f415cb305 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.499676] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 748.499927] env[62569]: INFO nova.compute.manager [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Took 6.08 seconds to spawn the instance on the hypervisor. [ 748.500099] env[62569]: DEBUG nova.compute.manager [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 748.500376] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 015140aa-8336-40e2-856c-723277a48f6e/015140aa-8336-40e2-856c-723277a48f6e.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 748.501190] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 748.501758] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ec9a8e3-bf7f-415b-818b-3a33ad21ee02 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.504472] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9d4f22ee-2b37-4231-b0a7-f7abcf58fb30 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.512290] env[62569]: DEBUG nova.network.neutron [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.519462] env[62569]: INFO nova.scheduler.client.report [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Deleted allocations for instance 3551a2e8-2f7f-475d-9c5c-2c8e88335bc2 [ 748.534176] env[62569]: DEBUG nova.network.neutron [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.536524] env[62569]: DEBUG oslo_concurrency.lockutils [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.537308] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 748.537308] env[62569]: value = "task-1249914" [ 748.537308] env[62569]: _type = "Task" [ 748.537308] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.550351] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249914, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.620029] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Acquiring lock "fdeac8b8-a34d-4b4c-8f72-281f84e22ebe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.620328] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Lock "fdeac8b8-a34d-4b4c-8f72-281f84e22ebe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.846542] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a437646-f988-48a7-9015-d5f6b10995c7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.856653] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81a717f8-5cc9-4030-8f71-30fd0a789e57 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.889527] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58f3547c-dd82-476b-b554-34afd51fa39b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.899205] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50712738-9846-4068-b637-9b893e00659b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.912111] env[62569]: DEBUG nova.compute.provider_tree [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.984035] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Acquiring lock "3f504d5c-3431-4a80-a7d9-e52af01cf24e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.984336] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Lock "3f504d5c-3431-4a80-a7d9-e52af01cf24e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.984550] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Acquiring lock "3f504d5c-3431-4a80-a7d9-e52af01cf24e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.984733] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Lock "3f504d5c-3431-4a80-a7d9-e52af01cf24e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.984896] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Lock "3f504d5c-3431-4a80-a7d9-e52af01cf24e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.987059] env[62569]: INFO nova.compute.manager [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Terminating instance [ 749.042676] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8cdf74c6-6497-4437-8418-c3a603770bbb tempest-ServersV294TestFqdnHostnames-1866092583 tempest-ServersV294TestFqdnHostnames-1866092583-project-member] Lock "3551a2e8-2f7f-475d-9c5c-2c8e88335bc2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 159.020s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.044074] env[62569]: INFO nova.compute.manager [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] [instance: 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda] Took 0.58 seconds to deallocate network for instance. [ 749.057827] env[62569]: INFO nova.compute.manager [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Took 30.58 seconds to build instance. [ 749.066133] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249914, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088007} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.068413] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 749.068947] env[62569]: INFO nova.compute.manager [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Took 26.30 seconds to build instance. [ 749.070272] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fda51072-0f5e-4456-ad9e-fde9bc3c4c20 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.093014] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 015140aa-8336-40e2-856c-723277a48f6e/015140aa-8336-40e2-856c-723277a48f6e.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 749.093654] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b60a0b29-bbd8-41c9-9cde-f472b578b442 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.115844] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 749.115844] env[62569]: value = "task-1249915" [ 749.115844] env[62569]: _type = "Task" [ 749.115844] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.127211] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249915, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.183057] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "b47cd2d7-0cd2-41af-8ed1-a6dfca323516" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.183057] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "b47cd2d7-0cd2-41af-8ed1-a6dfca323516" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.415641] env[62569]: DEBUG nova.scheduler.client.report [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 749.491373] env[62569]: DEBUG nova.compute.manager [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 749.491373] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 749.492797] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1c1b78-539b-465a-8482-17d8982bd5b1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.501737] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 749.502556] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c69d92ba-fa30-465b-9cdf-e3fe4b4539bb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.509889] env[62569]: DEBUG oslo_vmware.api [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Waiting for the task: (returnval){ [ 749.509889] env[62569]: value = "task-1249916" [ 749.509889] env[62569]: _type = "Task" [ 749.509889] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.520049] env[62569]: DEBUG oslo_vmware.api [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': task-1249916, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.540941] env[62569]: DEBUG nova.compute.manager [req-705f6028-ef82-425e-934e-a607513613bd req-d5eccb3d-c48c-48fa-aabb-05176fba4b1b service nova] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Received event network-changed-3da67b46-e89d-4a96-8dd9-752956b9b270 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 749.541100] env[62569]: DEBUG nova.compute.manager [req-705f6028-ef82-425e-934e-a607513613bd req-d5eccb3d-c48c-48fa-aabb-05176fba4b1b service nova] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Refreshing instance network info cache due to event network-changed-3da67b46-e89d-4a96-8dd9-752956b9b270. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 749.541322] env[62569]: DEBUG oslo_concurrency.lockutils [req-705f6028-ef82-425e-934e-a607513613bd req-d5eccb3d-c48c-48fa-aabb-05176fba4b1b service nova] Acquiring lock "refresh_cache-a0b406c3-9466-41bd-9de1-e675cab2ceef" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.541464] env[62569]: DEBUG oslo_concurrency.lockutils [req-705f6028-ef82-425e-934e-a607513613bd req-d5eccb3d-c48c-48fa-aabb-05176fba4b1b service nova] Acquired lock "refresh_cache-a0b406c3-9466-41bd-9de1-e675cab2ceef" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.541620] env[62569]: DEBUG nova.network.neutron [req-705f6028-ef82-425e-934e-a607513613bd req-d5eccb3d-c48c-48fa-aabb-05176fba4b1b service nova] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Refreshing network info cache for port 3da67b46-e89d-4a96-8dd9-752956b9b270 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 749.555656] env[62569]: DEBUG nova.compute.manager [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 749.558838] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e0a1d0f3-a86a-4420-8ebc-475af2de7eab tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "a0b406c3-9466-41bd-9de1-e675cab2ceef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.864s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.575465] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c3493402-4a8a-4a0e-9608-64bc8f892e86 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lock "94392d48-223d-4205-9627-25488a468769" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.645s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.626486] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249915, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.923441] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.189s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.924938] env[62569]: DEBUG nova.compute.manager [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 749.927167] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.033s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.929164] env[62569]: INFO nova.compute.claims [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 750.024530] env[62569]: DEBUG oslo_vmware.api [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': task-1249916, 'name': PowerOffVM_Task, 'duration_secs': 0.242223} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.024944] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 750.025229] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 750.025563] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bf2ca33b-d130-4f0e-bc5b-005fa8344234 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.061135] env[62569]: DEBUG nova.compute.manager [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 750.079183] env[62569]: DEBUG nova.compute.manager [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 750.083254] env[62569]: DEBUG oslo_concurrency.lockutils [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.085056] env[62569]: INFO nova.scheduler.client.report [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Deleted allocations for instance 0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda [ 750.096627] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 750.096855] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 750.097070] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Deleting the datastore file [datastore2] 3f504d5c-3431-4a80-a7d9-e52af01cf24e {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 750.097661] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e9e9c5f6-17fe-4467-b14d-503c71665e08 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.109666] env[62569]: DEBUG oslo_vmware.api [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Waiting for the task: (returnval){ [ 750.109666] env[62569]: value = "task-1249918" [ 750.109666] env[62569]: _type = "Task" [ 750.109666] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.119250] env[62569]: DEBUG oslo_vmware.api [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': task-1249918, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.133744] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249915, 'name': ReconfigVM_Task, 'duration_secs': 0.710022} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.134265] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 015140aa-8336-40e2-856c-723277a48f6e/015140aa-8336-40e2-856c-723277a48f6e.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 750.136018] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fbac0a54-f0b4-4282-886a-1d2e063e2272 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.144354] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 750.144354] env[62569]: value = "task-1249919" [ 750.144354] env[62569]: _type = "Task" [ 750.144354] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.154736] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249919, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.434128] env[62569]: DEBUG nova.compute.utils [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 750.443624] env[62569]: DEBUG nova.compute.manager [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 750.443624] env[62569]: DEBUG nova.network.neutron [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 750.511653] env[62569]: DEBUG nova.network.neutron [req-705f6028-ef82-425e-934e-a607513613bd req-d5eccb3d-c48c-48fa-aabb-05176fba4b1b service nova] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Updated VIF entry in instance network info cache for port 3da67b46-e89d-4a96-8dd9-752956b9b270. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 750.512047] env[62569]: DEBUG nova.network.neutron [req-705f6028-ef82-425e-934e-a607513613bd req-d5eccb3d-c48c-48fa-aabb-05176fba4b1b service nova] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Updating instance_info_cache with network_info: [{"id": "3da67b46-e89d-4a96-8dd9-752956b9b270", "address": "fa:16:3e:f6:2c:c2", "network": {"id": "ba6b71b1-bd7b-493e-8682-68d842fa935f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1892479706-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e9ea55b00b574a7db0530efcb498ca2b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3da67b46-e8", "ovs_interfaceid": "3da67b46-e89d-4a96-8dd9-752956b9b270", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.516038] env[62569]: DEBUG nova.policy [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7f4cb85930884f69aeff1ebbf6abd775', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ddafa4837ce64bf9aec427795e5f48a2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 750.585903] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.597093] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12d74d36-f3a9-4276-a70f-59f6b75d2a6b tempest-AttachInterfacesV270Test-588430262 tempest-AttachInterfacesV270Test-588430262-project-member] Lock "0b3b6f33-b90f-4c3e-bdb7-7d67b7793bda" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 154.083s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.614173] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.624814] env[62569]: DEBUG oslo_vmware.api [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Task: {'id': task-1249918, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.389536} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.625133] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 750.625354] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 750.625501] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 750.625668] env[62569]: INFO nova.compute.manager [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Took 1.13 seconds to destroy the instance on the hypervisor. [ 750.625906] env[62569]: DEBUG oslo.service.loopingcall [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 750.626728] env[62569]: DEBUG nova.compute.manager [-] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 750.626841] env[62569]: DEBUG nova.network.neutron [-] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 750.656136] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249919, 'name': Rename_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.948383] env[62569]: DEBUG nova.compute.manager [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 750.967065] env[62569]: DEBUG nova.network.neutron [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Successfully created port: dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 751.015646] env[62569]: DEBUG oslo_concurrency.lockutils [req-705f6028-ef82-425e-934e-a607513613bd req-d5eccb3d-c48c-48fa-aabb-05176fba4b1b service nova] Releasing lock "refresh_cache-a0b406c3-9466-41bd-9de1-e675cab2ceef" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.102353] env[62569]: DEBUG nova.compute.manager [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 751.160203] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249919, 'name': Rename_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.319729] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45c45558-7df3-4255-9da0-6982100bcf9b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.329814] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a9a4d76-0351-4fd4-8f41-59752a7d05b7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.371690] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-358259a4-b309-4c7d-8c44-eff0a6e9de70 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.380172] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9f6264b-03e1-4a27-a7ad-583b9fd12195 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.394474] env[62569]: DEBUG nova.compute.provider_tree [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.550297] env[62569]: DEBUG nova.network.neutron [-] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.623329] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.657665] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249919, 'name': Rename_Task, 'duration_secs': 1.191665} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.657965] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 751.658266] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c10a7d8e-c275-402c-b354-31de9bc73690 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.666039] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 751.666039] env[62569]: value = "task-1249920" [ 751.666039] env[62569]: _type = "Task" [ 751.666039] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.674939] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249920, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.740014] env[62569]: DEBUG nova.compute.manager [req-0ebd3688-cc97-46e1-938b-bdb436144073 req-ccc203cc-88f1-4456-9cc8-dba6ccd7fcca service nova] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Received event network-vif-deleted-b0b6ffe7-e392-40a8-b423-566a269133c2 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 751.897992] env[62569]: DEBUG nova.scheduler.client.report [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 751.965016] env[62569]: DEBUG nova.compute.manager [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 751.995484] env[62569]: DEBUG nova.virt.hardware [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 751.995643] env[62569]: DEBUG nova.virt.hardware [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 751.999088] env[62569]: DEBUG nova.virt.hardware [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 751.999088] env[62569]: DEBUG nova.virt.hardware [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 751.999088] env[62569]: DEBUG nova.virt.hardware [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 751.999088] env[62569]: DEBUG nova.virt.hardware [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 751.999088] env[62569]: DEBUG nova.virt.hardware [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 751.999288] env[62569]: DEBUG nova.virt.hardware [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 751.999288] env[62569]: DEBUG nova.virt.hardware [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 751.999288] env[62569]: DEBUG nova.virt.hardware [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 751.999288] env[62569]: DEBUG nova.virt.hardware [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 751.999288] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-828a3f24-10a2-46a1-98d1-fdbc294bfc75 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.008159] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-817d9fdb-1b7c-47ce-86da-4d0066f97544 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.052698] env[62569]: INFO nova.compute.manager [-] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Took 1.43 seconds to deallocate network for instance. [ 752.179949] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249920, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.408272] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.479s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.408272] env[62569]: DEBUG nova.compute.manager [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 752.410851] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.363s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.560148] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.681832] env[62569]: DEBUG oslo_vmware.api [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249920, 'name': PowerOnVM_Task, 'duration_secs': 0.749818} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.684180] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 752.684412] env[62569]: INFO nova.compute.manager [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Took 7.89 seconds to spawn the instance on the hypervisor. [ 752.684601] env[62569]: DEBUG nova.compute.manager [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 752.685891] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3087ba43-90d7-4112-9bd4-d51428f9e810 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.764653] env[62569]: DEBUG nova.network.neutron [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Successfully updated port: dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 752.917198] env[62569]: DEBUG nova.compute.utils [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 752.922877] env[62569]: DEBUG nova.compute.manager [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Not allocating networking since 'none' was specified. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 753.209613] env[62569]: INFO nova.compute.manager [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Took 28.57 seconds to build instance. [ 753.268042] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Acquiring lock "refresh_cache-c515d85e-bcb5-4bac-bacb-1e558f38171f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.268042] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Acquired lock "refresh_cache-c515d85e-bcb5-4bac-bacb-1e558f38171f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.268042] env[62569]: DEBUG nova.network.neutron [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 753.424125] env[62569]: DEBUG nova.compute.manager [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 753.449486] env[62569]: WARNING nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 93c8445d-689c-4f77-836f-95eafafbb6a2 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 753.449740] env[62569]: WARNING nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 3f504d5c-3431-4a80-a7d9-e52af01cf24e is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 753.449811] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance a0b406c3-9466-41bd-9de1-e675cab2ceef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 753.449898] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 94392d48-223d-4205-9627-25488a468769 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 753.450029] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 015140aa-8336-40e2-856c-723277a48f6e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 753.450154] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance c515d85e-bcb5-4bac-bacb-1e558f38171f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 753.450328] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance fd763e12-5b29-48f8-8256-cb6205e0d119 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 753.607191] env[62569]: INFO nova.compute.manager [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Rebuilding instance [ 753.647975] env[62569]: DEBUG nova.compute.manager [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 753.648826] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-726ba53e-54ea-4425-a907-9c6fa690d59d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.713628] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3de93cd9-b345-4a19-ba07-ae2f749015b9 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lock "015140aa-8336-40e2-856c-723277a48f6e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.586s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.820882] env[62569]: DEBUG nova.network.neutron [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.838412] env[62569]: DEBUG nova.compute.manager [req-e8ea31e5-c2c4-4de2-b73c-9399f91ce852 req-ea0b510e-6858-499e-8433-bc6ae45ac6f6 service nova] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Received event network-vif-plugged-dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 753.838628] env[62569]: DEBUG oslo_concurrency.lockutils [req-e8ea31e5-c2c4-4de2-b73c-9399f91ce852 req-ea0b510e-6858-499e-8433-bc6ae45ac6f6 service nova] Acquiring lock "c515d85e-bcb5-4bac-bacb-1e558f38171f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.838835] env[62569]: DEBUG oslo_concurrency.lockutils [req-e8ea31e5-c2c4-4de2-b73c-9399f91ce852 req-ea0b510e-6858-499e-8433-bc6ae45ac6f6 service nova] Lock "c515d85e-bcb5-4bac-bacb-1e558f38171f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.839109] env[62569]: DEBUG oslo_concurrency.lockutils [req-e8ea31e5-c2c4-4de2-b73c-9399f91ce852 req-ea0b510e-6858-499e-8433-bc6ae45ac6f6 service nova] Lock "c515d85e-bcb5-4bac-bacb-1e558f38171f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.839175] env[62569]: DEBUG nova.compute.manager [req-e8ea31e5-c2c4-4de2-b73c-9399f91ce852 req-ea0b510e-6858-499e-8433-bc6ae45ac6f6 service nova] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] No waiting events found dispatching network-vif-plugged-dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 753.839365] env[62569]: WARNING nova.compute.manager [req-e8ea31e5-c2c4-4de2-b73c-9399f91ce852 req-ea0b510e-6858-499e-8433-bc6ae45ac6f6 service nova] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Received unexpected event network-vif-plugged-dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897 for instance with vm_state building and task_state spawning. [ 753.839517] env[62569]: DEBUG nova.compute.manager [req-e8ea31e5-c2c4-4de2-b73c-9399f91ce852 req-ea0b510e-6858-499e-8433-bc6ae45ac6f6 service nova] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Received event network-changed-dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 753.839675] env[62569]: DEBUG nova.compute.manager [req-e8ea31e5-c2c4-4de2-b73c-9399f91ce852 req-ea0b510e-6858-499e-8433-bc6ae45ac6f6 service nova] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Refreshing instance network info cache due to event network-changed-dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 753.839839] env[62569]: DEBUG oslo_concurrency.lockutils [req-e8ea31e5-c2c4-4de2-b73c-9399f91ce852 req-ea0b510e-6858-499e-8433-bc6ae45ac6f6 service nova] Acquiring lock "refresh_cache-c515d85e-bcb5-4bac-bacb-1e558f38171f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.954170] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 77a1b192-6aff-4fee-93d7-57cebcdce626 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 754.059015] env[62569]: DEBUG nova.network.neutron [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Updating instance_info_cache with network_info: [{"id": "dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897", "address": "fa:16:3e:6d:4d:fa", "network": {"id": "6135c547-3ec2-4442-8616-2dc3970b99b6", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-2050038244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ddafa4837ce64bf9aec427795e5f48a2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fc48e29b-113c-4849-850c-35435eab4052", "external-id": "nsx-vlan-transportzone-958", "segmentation_id": 958, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdedebbd2-90", "ovs_interfaceid": "dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.215274] env[62569]: DEBUG nova.compute.manager [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 754.443702] env[62569]: DEBUG nova.compute.manager [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 754.457073] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 22094c32-5f50-4f86-a77b-cd4adcf8998a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 754.476941] env[62569]: DEBUG nova.virt.hardware [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 754.477293] env[62569]: DEBUG nova.virt.hardware [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 754.477525] env[62569]: DEBUG nova.virt.hardware [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 754.477771] env[62569]: DEBUG nova.virt.hardware [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 754.478013] env[62569]: DEBUG nova.virt.hardware [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 754.478185] env[62569]: DEBUG nova.virt.hardware [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 754.478402] env[62569]: DEBUG nova.virt.hardware [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 754.478566] env[62569]: DEBUG nova.virt.hardware [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 754.478740] env[62569]: DEBUG nova.virt.hardware [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 754.478941] env[62569]: DEBUG nova.virt.hardware [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 754.479142] env[62569]: DEBUG nova.virt.hardware [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 754.480345] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30fe0f4b-493f-46a8-86e4-967e9e700094 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.490549] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae32ed66-7457-4fb4-b0d6-c51bfd19488f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.504674] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Instance VIF info [] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 754.512040] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Creating folder: Project (424cbe0b3f124b78b411ab8d297ad88d). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 754.512416] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ffbc8e21-052d-4b14-84d3-8c97d900b8a6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.522873] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Created folder: Project (424cbe0b3f124b78b411ab8d297ad88d) in parent group-v269330. [ 754.523079] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Creating folder: Instances. Parent ref: group-v269364. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 754.523314] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ec6521e3-066b-47a8-b9bf-f9718cb3a44c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.532446] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Created folder: Instances in parent group-v269364. [ 754.532677] env[62569]: DEBUG oslo.service.loopingcall [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 754.532883] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 754.533096] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4c0f3253-d083-42fe-90f2-89a02b53ab08 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.549135] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 754.549135] env[62569]: value = "task-1249923" [ 754.549135] env[62569]: _type = "Task" [ 754.549135] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.557632] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249923, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.561265] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Releasing lock "refresh_cache-c515d85e-bcb5-4bac-bacb-1e558f38171f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.561557] env[62569]: DEBUG nova.compute.manager [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Instance network_info: |[{"id": "dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897", "address": "fa:16:3e:6d:4d:fa", "network": {"id": "6135c547-3ec2-4442-8616-2dc3970b99b6", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-2050038244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ddafa4837ce64bf9aec427795e5f48a2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fc48e29b-113c-4849-850c-35435eab4052", "external-id": "nsx-vlan-transportzone-958", "segmentation_id": 958, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdedebbd2-90", "ovs_interfaceid": "dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 754.561955] env[62569]: DEBUG oslo_concurrency.lockutils [req-e8ea31e5-c2c4-4de2-b73c-9399f91ce852 req-ea0b510e-6858-499e-8433-bc6ae45ac6f6 service nova] Acquired lock "refresh_cache-c515d85e-bcb5-4bac-bacb-1e558f38171f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.562093] env[62569]: DEBUG nova.network.neutron [req-e8ea31e5-c2c4-4de2-b73c-9399f91ce852 req-ea0b510e-6858-499e-8433-bc6ae45ac6f6 service nova] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Refreshing network info cache for port dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 754.563176] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:4d:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fc48e29b-113c-4849-850c-35435eab4052', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 754.571216] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Creating folder: Project (ddafa4837ce64bf9aec427795e5f48a2). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 754.571425] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f0479dfa-afeb-4bf9-9b3a-4227ad42fc36 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.586101] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Created folder: Project (ddafa4837ce64bf9aec427795e5f48a2) in parent group-v269330. [ 754.586411] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Creating folder: Instances. Parent ref: group-v269367. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 754.586524] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8b427a91-21e8-4c25-ac48-690ba3527d3f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.595152] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Created folder: Instances in parent group-v269367. [ 754.595425] env[62569]: DEBUG oslo.service.loopingcall [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 754.595586] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 754.595794] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-02d96b6f-d017-41a4-bc2d-1e8dba8ccacd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.619689] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 754.619689] env[62569]: value = "task-1249926" [ 754.619689] env[62569]: _type = "Task" [ 754.619689] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.628018] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249926, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.662472] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 754.662809] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2ee7a4fa-be7c-4979-89bd-0ec3a4304abc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.670224] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 754.670224] env[62569]: value = "task-1249927" [ 754.670224] env[62569]: _type = "Task" [ 754.670224] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.680310] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249927, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.767749] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.962189] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 4feb7b78-9f7c-4e64-b0a7-870ed73adf97 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 755.059887] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249923, 'name': CreateVM_Task, 'duration_secs': 0.297862} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.060420] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 755.061080] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.061350] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.061715] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 755.062274] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d60b33f2-7891-4a2f-8226-09733b1ea807 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.067669] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Waiting for the task: (returnval){ [ 755.067669] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527b5bf3-0f02-ec4d-26aa-8fe69bee8f97" [ 755.067669] env[62569]: _type = "Task" [ 755.067669] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.080645] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527b5bf3-0f02-ec4d-26aa-8fe69bee8f97, 'name': SearchDatastore_Task, 'duration_secs': 0.009886} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.080911] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.081208] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 755.081534] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.081739] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.081976] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 755.082336] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d8cdfb99-411d-4b5f-96cd-0c2df8666c97 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.090750] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 755.090750] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 755.094263] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6db43726-28e2-406e-9f64-7da8f5e1575a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.101773] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Waiting for the task: (returnval){ [ 755.101773] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5268ce37-d058-e121-cd4b-d28acb33bd00" [ 755.101773] env[62569]: _type = "Task" [ 755.101773] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.110372] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5268ce37-d058-e121-cd4b-d28acb33bd00, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.128828] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249926, 'name': CreateVM_Task, 'duration_secs': 0.326221} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.129391] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 755.129972] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.130164] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.130496] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 755.130750] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eefbbf5f-e493-4c4b-99d4-00c8b1f41c73 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.135158] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Waiting for the task: (returnval){ [ 755.135158] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52198912-22dd-587a-9ffa-585614ccfc99" [ 755.135158] env[62569]: _type = "Task" [ 755.135158] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.142762] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52198912-22dd-587a-9ffa-585614ccfc99, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.179123] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249927, 'name': PowerOffVM_Task, 'duration_secs': 0.133531} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.182266] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 755.182575] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 755.183934] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a2a11a8-50c9-4728-b104-3c46e82c512d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.191248] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 755.192041] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-64f9c8b0-d2f9-42cb-adb1-a5815adec8e0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.219717] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 755.219717] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 755.219717] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Deleting the datastore file [datastore1] 015140aa-8336-40e2-856c-723277a48f6e {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 755.219717] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e417839f-8cea-4a3a-b82a-2791548a690f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.225650] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 755.225650] env[62569]: value = "task-1249929" [ 755.225650] env[62569]: _type = "Task" [ 755.225650] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.235170] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249929, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.446018] env[62569]: DEBUG nova.network.neutron [req-e8ea31e5-c2c4-4de2-b73c-9399f91ce852 req-ea0b510e-6858-499e-8433-bc6ae45ac6f6 service nova] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Updated VIF entry in instance network info cache for port dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 755.446018] env[62569]: DEBUG nova.network.neutron [req-e8ea31e5-c2c4-4de2-b73c-9399f91ce852 req-ea0b510e-6858-499e-8433-bc6ae45ac6f6 service nova] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Updating instance_info_cache with network_info: [{"id": "dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897", "address": "fa:16:3e:6d:4d:fa", "network": {"id": "6135c547-3ec2-4442-8616-2dc3970b99b6", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-2050038244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ddafa4837ce64bf9aec427795e5f48a2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fc48e29b-113c-4849-850c-35435eab4052", "external-id": "nsx-vlan-transportzone-958", "segmentation_id": 958, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdedebbd2-90", "ovs_interfaceid": "dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.465401] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 755.614980] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5268ce37-d058-e121-cd4b-d28acb33bd00, 'name': SearchDatastore_Task, 'duration_secs': 0.012872} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.615644] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-599b1418-eb9f-4b13-8879-087defcb0db4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.621256] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Waiting for the task: (returnval){ [ 755.621256] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5233bb4c-fed6-0306-c764-7f78762b55cd" [ 755.621256] env[62569]: _type = "Task" [ 755.621256] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.629178] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5233bb4c-fed6-0306-c764-7f78762b55cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.644869] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52198912-22dd-587a-9ffa-585614ccfc99, 'name': SearchDatastore_Task, 'duration_secs': 0.010484} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.645296] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.645709] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 755.645770] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.735734] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249929, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.338075} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.735984] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 755.736177] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 755.736370] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 755.947833] env[62569]: DEBUG oslo_concurrency.lockutils [req-e8ea31e5-c2c4-4de2-b73c-9399f91ce852 req-ea0b510e-6858-499e-8433-bc6ae45ac6f6 service nova] Releasing lock "refresh_cache-c515d85e-bcb5-4bac-bacb-1e558f38171f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.973023] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance b6e5eefc-8c06-445b-a3af-9404578b6179 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 756.137844] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5233bb4c-fed6-0306-c764-7f78762b55cd, 'name': SearchDatastore_Task, 'duration_secs': 0.014067} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.138293] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.138578] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] fd763e12-5b29-48f8-8256-cb6205e0d119/fd763e12-5b29-48f8-8256-cb6205e0d119.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 756.139168] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.139414] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 756.139780] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8bcc3bd9-8711-493d-9571-51a3e4cd1db5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.142921] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4c3875fa-b003-4530-ba3e-e2a3c44fdf94 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.151484] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Waiting for the task: (returnval){ [ 756.151484] env[62569]: value = "task-1249930" [ 756.151484] env[62569]: _type = "Task" [ 756.151484] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.158761] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 756.158994] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 756.160340] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02f741ca-c01c-4504-8363-3d8ebbbce970 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.166178] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Task: {'id': task-1249930, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.169451] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Waiting for the task: (returnval){ [ 756.169451] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a98726-6031-61da-403d-73a2ce6101e8" [ 756.169451] env[62569]: _type = "Task" [ 756.169451] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.177183] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a98726-6031-61da-403d-73a2ce6101e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.478065] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 37a4c60a-e660-4c91-bcfe-72638a4397b0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 756.662024] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Task: {'id': task-1249930, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.686021] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a98726-6031-61da-403d-73a2ce6101e8, 'name': SearchDatastore_Task, 'duration_secs': 0.024775} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.686021] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f91856c-a155-4319-ad5b-a454dd9c4f96 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.690274] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Waiting for the task: (returnval){ [ 756.690274] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52c06e79-391e-ab9f-b822-44f0e1061f9d" [ 756.690274] env[62569]: _type = "Task" [ 756.690274] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.699782] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52c06e79-391e-ab9f-b822-44f0e1061f9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.780538] env[62569]: DEBUG nova.virt.hardware [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 756.782028] env[62569]: DEBUG nova.virt.hardware [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 756.782123] env[62569]: DEBUG nova.virt.hardware [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 756.782324] env[62569]: DEBUG nova.virt.hardware [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 756.782477] env[62569]: DEBUG nova.virt.hardware [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 756.782624] env[62569]: DEBUG nova.virt.hardware [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 756.782835] env[62569]: DEBUG nova.virt.hardware [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 756.782998] env[62569]: DEBUG nova.virt.hardware [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 756.783201] env[62569]: DEBUG nova.virt.hardware [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 756.783372] env[62569]: DEBUG nova.virt.hardware [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 756.783628] env[62569]: DEBUG nova.virt.hardware [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 756.784530] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144ce278-d64b-4433-bcb4-cf7382af2c5f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.797300] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59a79329-a2df-4c1c-b03f-d56d28fe5719 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.817223] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Instance VIF info [] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 756.827279] env[62569]: DEBUG oslo.service.loopingcall [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 756.827279] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 756.827279] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-570b6ade-0c3d-47b6-9c93-81c360e11b72 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.846319] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 756.846319] env[62569]: value = "task-1249931" [ 756.846319] env[62569]: _type = "Task" [ 756.846319] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.854364] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249931, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.981692] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 06067572-a9fd-43a4-91cd-383a62f53885 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 757.165491] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Task: {'id': task-1249930, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.56002} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.165800] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] fd763e12-5b29-48f8-8256-cb6205e0d119/fd763e12-5b29-48f8-8256-cb6205e0d119.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 757.166120] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 757.166461] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d6f3e9c1-6549-4b7d-8870-eb1015d25bab {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.178423] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Waiting for the task: (returnval){ [ 757.178423] env[62569]: value = "task-1249932" [ 757.178423] env[62569]: _type = "Task" [ 757.178423] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.190793] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Task: {'id': task-1249932, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.200112] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52c06e79-391e-ab9f-b822-44f0e1061f9d, 'name': SearchDatastore_Task, 'duration_secs': 0.010019} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.200559] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.200916] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] c515d85e-bcb5-4bac-bacb-1e558f38171f/c515d85e-bcb5-4bac-bacb-1e558f38171f.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 757.201572] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-abd74c9c-c243-4e52-ad4a-6109d4a68ff3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.208295] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Waiting for the task: (returnval){ [ 757.208295] env[62569]: value = "task-1249933" [ 757.208295] env[62569]: _type = "Task" [ 757.208295] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.218046] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': task-1249933, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.358730] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249931, 'name': CreateVM_Task, 'duration_secs': 0.297317} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.358965] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 757.359552] env[62569]: DEBUG oslo_concurrency.lockutils [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.359858] env[62569]: DEBUG oslo_concurrency.lockutils [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.360244] env[62569]: DEBUG oslo_concurrency.lockutils [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 757.360522] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5adea695-6e0e-43ff-959c-0e1d3897d194 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.365608] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 757.365608] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52af50f9-3a2d-92c3-ca3a-ddbce7a9c6fc" [ 757.365608] env[62569]: _type = "Task" [ 757.365608] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.375555] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52af50f9-3a2d-92c3-ca3a-ddbce7a9c6fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.484875] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance b3ccc28b-68d0-461d-a67d-b5ad179a80f9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 757.688732] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Task: {'id': task-1249932, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074611} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.689141] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 757.690016] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08958153-951a-43de-b6dc-a27a12da9f3f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.713797] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] fd763e12-5b29-48f8-8256-cb6205e0d119/fd763e12-5b29-48f8-8256-cb6205e0d119.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 757.714282] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2866a113-c0f7-49be-8fdc-b36c3b4618f2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.742497] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': task-1249933, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.743952] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Waiting for the task: (returnval){ [ 757.743952] env[62569]: value = "task-1249934" [ 757.743952] env[62569]: _type = "Task" [ 757.743952] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.751134] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Task: {'id': task-1249934, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.877921] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52af50f9-3a2d-92c3-ca3a-ddbce7a9c6fc, 'name': SearchDatastore_Task, 'duration_secs': 0.022447} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.878295] env[62569]: DEBUG oslo_concurrency.lockutils [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.878534] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 757.878830] env[62569]: DEBUG oslo_concurrency.lockutils [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.878993] env[62569]: DEBUG oslo_concurrency.lockutils [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.879184] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 757.880186] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6ee1a6e0-72e3-4886-b27c-a898d37b7784 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.892391] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 757.892649] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 757.898412] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b28fff18-d1ea-4119-9d23-d768aa9cafc6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.907775] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "cb9d8991-b7ba-4bf7-9d22-1391f4cea768" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.911132] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cb9d8991-b7ba-4bf7-9d22-1391f4cea768" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.911997] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 757.911997] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52284c59-57f2-1f80-4c8f-76f1e9aa5bf0" [ 757.911997] env[62569]: _type = "Task" [ 757.911997] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.930951] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52284c59-57f2-1f80-4c8f-76f1e9aa5bf0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.988478] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 337d88a2-30b9-4846-929e-042bd7a64a65 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 758.226865] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': task-1249933, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.907516} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.227184] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] c515d85e-bcb5-4bac-bacb-1e558f38171f/c515d85e-bcb5-4bac-bacb-1e558f38171f.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 758.227410] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 758.227664] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6b7a1cd7-0050-4336-8182-38d4f3a80b4c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.245505] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Waiting for the task: (returnval){ [ 758.245505] env[62569]: value = "task-1249935" [ 758.245505] env[62569]: _type = "Task" [ 758.245505] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.256650] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': task-1249935, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.261228] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Task: {'id': task-1249934, 'name': ReconfigVM_Task, 'duration_secs': 0.309587} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.261477] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Reconfigured VM instance instance-00000038 to attach disk [datastore2] fd763e12-5b29-48f8-8256-cb6205e0d119/fd763e12-5b29-48f8-8256-cb6205e0d119.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 758.262432] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-923437cb-8a21-47c5-b769-3cc42d126cf0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.269843] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Waiting for the task: (returnval){ [ 758.269843] env[62569]: value = "task-1249936" [ 758.269843] env[62569]: _type = "Task" [ 758.269843] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.280583] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Task: {'id': task-1249936, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.422908] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52284c59-57f2-1f80-4c8f-76f1e9aa5bf0, 'name': SearchDatastore_Task, 'duration_secs': 0.023489} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.423744] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c39b4a5-62ee-4bee-bb87-4b31b713284d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.430696] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 758.430696] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529c875e-2c90-ff3b-917b-df675a6b365f" [ 758.430696] env[62569]: _type = "Task" [ 758.430696] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.438466] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529c875e-2c90-ff3b-917b-df675a6b365f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.491935] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance eb70341b-4282-4eca-b6a2-374db7c521c5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 758.732403] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.732675] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.757251] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': task-1249935, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079801} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.757461] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 758.758273] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3dde5e8-2057-4d49-89d4-e62e56181654 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.779336] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] c515d85e-bcb5-4bac-bacb-1e558f38171f/c515d85e-bcb5-4bac-bacb-1e558f38171f.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 758.782190] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5d1e86fd-1f9f-4b41-9db2-eb0a23e1a5be {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.801058] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Task: {'id': task-1249936, 'name': Rename_Task, 'duration_secs': 0.192013} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.802259] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 758.802557] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Waiting for the task: (returnval){ [ 758.802557] env[62569]: value = "task-1249937" [ 758.802557] env[62569]: _type = "Task" [ 758.802557] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.802743] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2f437404-8b02-4198-9d7b-d272ce8f8bb6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.813294] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': task-1249937, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.814436] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Waiting for the task: (returnval){ [ 758.814436] env[62569]: value = "task-1249938" [ 758.814436] env[62569]: _type = "Task" [ 758.814436] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.821173] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Task: {'id': task-1249938, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.940841] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529c875e-2c90-ff3b-917b-df675a6b365f, 'name': SearchDatastore_Task, 'duration_secs': 0.009244} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.941121] env[62569]: DEBUG oslo_concurrency.lockutils [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.941396] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 015140aa-8336-40e2-856c-723277a48f6e/015140aa-8336-40e2-856c-723277a48f6e.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 758.941773] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e298ecb7-ac81-46ce-9b50-032c0b577da2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.947807] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 758.947807] env[62569]: value = "task-1249939" [ 758.947807] env[62569]: _type = "Task" [ 758.947807] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.955773] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249939, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.995697] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance cefbca2e-609d-4954-bec6-52ffe095446f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 759.314400] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': task-1249937, 'name': ReconfigVM_Task, 'duration_secs': 0.481254} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.314749] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Reconfigured VM instance instance-00000037 to attach disk [datastore2] c515d85e-bcb5-4bac-bacb-1e558f38171f/c515d85e-bcb5-4bac-bacb-1e558f38171f.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 759.315529] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-554afb9a-3893-444a-a6f7-2be0330124a7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.325772] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Waiting for the task: (returnval){ [ 759.325772] env[62569]: value = "task-1249940" [ 759.325772] env[62569]: _type = "Task" [ 759.325772] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.329681] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Task: {'id': task-1249938, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.338334] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': task-1249940, 'name': Rename_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.457603] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249939, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474988} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.457887] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 015140aa-8336-40e2-856c-723277a48f6e/015140aa-8336-40e2-856c-723277a48f6e.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 759.458115] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 759.458365] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-45458c86-4fa1-4aee-accc-fe57697af928 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.463780] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 759.463780] env[62569]: value = "task-1249941" [ 759.463780] env[62569]: _type = "Task" [ 759.463780] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.471345] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249941, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.499373] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 807d3025-d6a7-4778-a829-a61e2c7495c9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 759.826396] env[62569]: DEBUG oslo_vmware.api [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Task: {'id': task-1249938, 'name': PowerOnVM_Task, 'duration_secs': 0.58563} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.826737] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 759.826737] env[62569]: INFO nova.compute.manager [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Took 5.38 seconds to spawn the instance on the hypervisor. [ 759.826846] env[62569]: DEBUG nova.compute.manager [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 759.827643] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29660df8-3317-4681-97a2-84b30e3708d1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.841633] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': task-1249940, 'name': Rename_Task, 'duration_secs': 0.334824} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.842133] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 759.842368] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6e2ac6f0-03a0-4e0f-922c-a4a1b17e6697 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.849135] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Waiting for the task: (returnval){ [ 759.849135] env[62569]: value = "task-1249942" [ 759.849135] env[62569]: _type = "Task" [ 759.849135] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.973689] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249941, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057143} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.974031] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 759.974886] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-123f12c5-4913-4370-a749-7781eccf31ce {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.996183] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 015140aa-8336-40e2-856c-723277a48f6e/015140aa-8336-40e2-856c-723277a48f6e.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 759.996509] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe25d503-bfba-443b-ab33-e94db3a47700 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.011844] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 390bcf25-689d-46ad-bffb-3670c3729397 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 760.018188] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 760.018188] env[62569]: value = "task-1249943" [ 760.018188] env[62569]: _type = "Task" [ 760.018188] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.026517] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249943, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.353034] env[62569]: INFO nova.compute.manager [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Took 24.49 seconds to build instance. [ 760.360644] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': task-1249942, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.514963] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance fdeac8b8-a34d-4b4c-8f72-281f84e22ebe has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 760.529463] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249943, 'name': ReconfigVM_Task, 'duration_secs': 0.283784} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.530368] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 015140aa-8336-40e2-856c-723277a48f6e/015140aa-8336-40e2-856c-723277a48f6e.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 760.530961] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5a30091e-a060-446e-a81e-4bbc72054d0f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.537117] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 760.537117] env[62569]: value = "task-1249944" [ 760.537117] env[62569]: _type = "Task" [ 760.537117] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.544521] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249944, 'name': Rename_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.860125] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f64ade1-f246-4912-9e58-5261ff1f017f tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Lock "fd763e12-5b29-48f8-8256-cb6205e0d119" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.641s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.860428] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': task-1249942, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.018542] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance b47cd2d7-0cd2-41af-8ed1-a6dfca323516 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 761.018811] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=62569) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 761.018960] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=62569) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 761.052989] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249944, 'name': Rename_Task, 'duration_secs': 0.127976} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.053707] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 761.053997] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-63758cea-9254-43f3-aa5f-db3f2e574216 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.061863] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 761.061863] env[62569]: value = "task-1249945" [ 761.061863] env[62569]: _type = "Task" [ 761.061863] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.072453] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249945, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.339336] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f27604a2-2f6f-4435-a5b5-62547bd030a3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.347342] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eff16ce-9911-40e0-b131-ba6d57c3048a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.357984] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': task-1249942, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.383832] env[62569]: DEBUG nova.compute.manager [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 761.387219] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64ea720d-cf1c-41fa-8860-a5cf0a657cc3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.395271] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92a514f0-93f2-4121-bd64-7ea82f1ff348 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.410401] env[62569]: DEBUG nova.compute.provider_tree [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 761.554561] env[62569]: DEBUG nova.compute.manager [None req-70472a41-07a0-4afa-b5bd-9790d4d6c179 tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 761.555597] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-682800ef-8d5e-4e15-90f3-cb028a6e2da5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.571959] env[62569]: DEBUG oslo_vmware.api [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249945, 'name': PowerOnVM_Task, 'duration_secs': 0.407966} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.572235] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 761.572434] env[62569]: DEBUG nova.compute.manager [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 761.573201] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f83a5d75-7a19-41bb-999d-d7ef1241f297 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.647370] env[62569]: DEBUG oslo_concurrency.lockutils [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Acquiring lock "fd763e12-5b29-48f8-8256-cb6205e0d119" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.647642] env[62569]: DEBUG oslo_concurrency.lockutils [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Lock "fd763e12-5b29-48f8-8256-cb6205e0d119" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.647853] env[62569]: DEBUG oslo_concurrency.lockutils [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Acquiring lock "fd763e12-5b29-48f8-8256-cb6205e0d119-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.648044] env[62569]: DEBUG oslo_concurrency.lockutils [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Lock "fd763e12-5b29-48f8-8256-cb6205e0d119-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.648390] env[62569]: DEBUG oslo_concurrency.lockutils [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Lock "fd763e12-5b29-48f8-8256-cb6205e0d119-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.653873] env[62569]: INFO nova.compute.manager [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Terminating instance [ 761.861687] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': task-1249942, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.913969] env[62569]: DEBUG nova.scheduler.client.report [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 761.992821] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.068859] env[62569]: INFO nova.compute.manager [None req-70472a41-07a0-4afa-b5bd-9790d4d6c179 tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] instance snapshotting [ 762.069474] env[62569]: DEBUG nova.objects.instance [None req-70472a41-07a0-4afa-b5bd-9790d4d6c179 tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Lazy-loading 'flavor' on Instance uuid fd763e12-5b29-48f8-8256-cb6205e0d119 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 762.087624] env[62569]: DEBUG oslo_concurrency.lockutils [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.157785] env[62569]: DEBUG oslo_concurrency.lockutils [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Acquiring lock "refresh_cache-fd763e12-5b29-48f8-8256-cb6205e0d119" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.157933] env[62569]: DEBUG oslo_concurrency.lockutils [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Acquired lock "refresh_cache-fd763e12-5b29-48f8-8256-cb6205e0d119" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.158172] env[62569]: DEBUG nova.network.neutron [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 762.361020] env[62569]: DEBUG oslo_vmware.api [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': task-1249942, 'name': PowerOnVM_Task, 'duration_secs': 2.034299} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.361020] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 762.361020] env[62569]: INFO nova.compute.manager [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Took 10.40 seconds to spawn the instance on the hypervisor. [ 762.361020] env[62569]: DEBUG nova.compute.manager [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 762.361285] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-283dab5f-3535-4a52-bd90-c92a17b9223b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.418112] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62569) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 762.418350] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.008s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.418709] env[62569]: DEBUG oslo_concurrency.lockutils [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.172s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.420161] env[62569]: INFO nova.compute.claims [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 762.422983] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 762.423149] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Cleaning up deleted instances {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11608}} [ 762.575855] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eb3b5cc-3b44-4f77-a4a0-eb843ce7c5a8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.594844] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f51a8851-d201-4705-ba84-d60d02b0653c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.679694] env[62569]: DEBUG nova.network.neutron [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.746866] env[62569]: DEBUG nova.network.neutron [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.884832] env[62569]: INFO nova.compute.manager [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Took 31.22 seconds to build instance. [ 762.908334] env[62569]: DEBUG oslo_concurrency.lockutils [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquiring lock "015140aa-8336-40e2-856c-723277a48f6e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.908642] env[62569]: DEBUG oslo_concurrency.lockutils [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lock "015140aa-8336-40e2-856c-723277a48f6e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.908868] env[62569]: DEBUG oslo_concurrency.lockutils [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquiring lock "015140aa-8336-40e2-856c-723277a48f6e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.909180] env[62569]: DEBUG oslo_concurrency.lockutils [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lock "015140aa-8336-40e2-856c-723277a48f6e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.909259] env[62569]: DEBUG oslo_concurrency.lockutils [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lock "015140aa-8336-40e2-856c-723277a48f6e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.911302] env[62569]: INFO nova.compute.manager [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Terminating instance [ 762.933808] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] There are 3 instances to clean {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11617}} [ 762.934234] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 4b9e4f73-4682-4170-b71e-1d3d1d860577] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 763.105197] env[62569]: DEBUG nova.compute.manager [None req-70472a41-07a0-4afa-b5bd-9790d4d6c179 tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Instance disappeared during snapshot {{(pid=62569) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4594}} [ 763.250493] env[62569]: DEBUG oslo_concurrency.lockutils [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Releasing lock "refresh_cache-fd763e12-5b29-48f8-8256-cb6205e0d119" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.250954] env[62569]: DEBUG nova.compute.manager [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 763.251185] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 763.252149] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00d8f62b-afac-4c01-bcc5-0fd687ca798a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.260771] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 763.260771] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2244206f-2bb8-448f-bddd-8ab92a90adb3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.267047] env[62569]: DEBUG oslo_vmware.api [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Waiting for the task: (returnval){ [ 763.267047] env[62569]: value = "task-1249946" [ 763.267047] env[62569]: _type = "Task" [ 763.267047] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.276576] env[62569]: DEBUG oslo_vmware.api [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Task: {'id': task-1249946, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.283417] env[62569]: DEBUG nova.compute.manager [None req-70472a41-07a0-4afa-b5bd-9790d4d6c179 tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Found 0 images (rotation: 2) {{(pid=62569) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4923}} [ 763.386908] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d05edf2-870f-4eb1-8ad4-9c1158309451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Lock "c515d85e-bcb5-4bac-bacb-1e558f38171f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.581s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.414433] env[62569]: DEBUG oslo_concurrency.lockutils [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquiring lock "refresh_cache-015140aa-8336-40e2-856c-723277a48f6e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.414620] env[62569]: DEBUG oslo_concurrency.lockutils [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquired lock "refresh_cache-015140aa-8336-40e2-856c-723277a48f6e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.414962] env[62569]: DEBUG nova.network.neutron [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 763.437506] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 6e35fe1c-05bc-4dcc-b7ec-e9998a647739] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 763.450615] env[62569]: DEBUG nova.scheduler.client.report [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Refreshing inventories for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 763.468662] env[62569]: DEBUG nova.scheduler.client.report [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Updating ProviderTree inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 763.468662] env[62569]: DEBUG nova.compute.provider_tree [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 763.479779] env[62569]: DEBUG nova.scheduler.client.report [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Refreshing aggregate associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, aggregates: None {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 763.499694] env[62569]: DEBUG nova.scheduler.client.report [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Refreshing trait associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 763.632598] env[62569]: DEBUG nova.compute.manager [req-d5d6f310-509b-48c1-afa7-adfce9e8ec50 req-1620420b-bd86-4d5c-9727-7cdb01a9e5fb service nova] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Received event network-changed-dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 763.632598] env[62569]: DEBUG nova.compute.manager [req-d5d6f310-509b-48c1-afa7-adfce9e8ec50 req-1620420b-bd86-4d5c-9727-7cdb01a9e5fb service nova] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Refreshing instance network info cache due to event network-changed-dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 763.632598] env[62569]: DEBUG oslo_concurrency.lockutils [req-d5d6f310-509b-48c1-afa7-adfce9e8ec50 req-1620420b-bd86-4d5c-9727-7cdb01a9e5fb service nova] Acquiring lock "refresh_cache-c515d85e-bcb5-4bac-bacb-1e558f38171f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.632598] env[62569]: DEBUG oslo_concurrency.lockutils [req-d5d6f310-509b-48c1-afa7-adfce9e8ec50 req-1620420b-bd86-4d5c-9727-7cdb01a9e5fb service nova] Acquired lock "refresh_cache-c515d85e-bcb5-4bac-bacb-1e558f38171f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.632598] env[62569]: DEBUG nova.network.neutron [req-d5d6f310-509b-48c1-afa7-adfce9e8ec50 req-1620420b-bd86-4d5c-9727-7cdb01a9e5fb service nova] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Refreshing network info cache for port dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 763.778441] env[62569]: DEBUG oslo_vmware.api [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Task: {'id': task-1249946, 'name': PowerOffVM_Task, 'duration_secs': 0.283995} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.778898] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 763.779189] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 763.779539] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-267ef011-7746-41d9-be80-dfb1e42b0d04 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.800269] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2303c60-a73f-4602-8da1-dcb6cb6e612f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.806048] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 763.806048] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 763.806048] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Deleting the datastore file [datastore2] fd763e12-5b29-48f8-8256-cb6205e0d119 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 763.807795] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0e6e39b9-5f01-4488-9d51-4cf0a3b217b9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.810501] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f22b6360-5719-439c-b549-8a956dcb25c0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.818014] env[62569]: DEBUG oslo_vmware.api [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Waiting for the task: (returnval){ [ 763.818014] env[62569]: value = "task-1249948" [ 763.818014] env[62569]: _type = "Task" [ 763.818014] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.847493] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3732704-2536-4288-96d4-0d7e0a3e4136 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.852960] env[62569]: DEBUG oslo_vmware.api [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Task: {'id': task-1249948, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.857810] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c448f5c-3be1-4ec4-91fb-40488ac0c210 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.871496] env[62569]: DEBUG nova.compute.provider_tree [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.889891] env[62569]: DEBUG nova.compute.manager [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 763.942026] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 17b0789b-6ffe-4e4e-8e4c-1cff2cc1d007] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 764.105803] env[62569]: DEBUG nova.network.neutron [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.187134] env[62569]: DEBUG nova.network.neutron [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.265770] env[62569]: DEBUG oslo_concurrency.lockutils [None req-23641bfb-0496-40f4-86f8-38994a4d9447 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Acquiring lock "c515d85e-bcb5-4bac-bacb-1e558f38171f" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.266015] env[62569]: DEBUG oslo_concurrency.lockutils [None req-23641bfb-0496-40f4-86f8-38994a4d9447 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Lock "c515d85e-bcb5-4bac-bacb-1e558f38171f" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.266265] env[62569]: INFO nova.compute.manager [None req-23641bfb-0496-40f4-86f8-38994a4d9447 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Rebooting instance [ 764.330638] env[62569]: DEBUG oslo_vmware.api [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Task: {'id': task-1249948, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131725} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.333364] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 764.333564] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 764.333743] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 764.333915] env[62569]: INFO nova.compute.manager [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Took 1.08 seconds to destroy the instance on the hypervisor. [ 764.334170] env[62569]: DEBUG oslo.service.loopingcall [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 764.334372] env[62569]: DEBUG nova.compute.manager [-] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 764.334469] env[62569]: DEBUG nova.network.neutron [-] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 764.363048] env[62569]: DEBUG nova.network.neutron [-] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.374310] env[62569]: DEBUG nova.scheduler.client.report [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 764.386697] env[62569]: DEBUG nova.network.neutron [req-d5d6f310-509b-48c1-afa7-adfce9e8ec50 req-1620420b-bd86-4d5c-9727-7cdb01a9e5fb service nova] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Updated VIF entry in instance network info cache for port dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 764.387063] env[62569]: DEBUG nova.network.neutron [req-d5d6f310-509b-48c1-afa7-adfce9e8ec50 req-1620420b-bd86-4d5c-9727-7cdb01a9e5fb service nova] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Updating instance_info_cache with network_info: [{"id": "dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897", "address": "fa:16:3e:6d:4d:fa", "network": {"id": "6135c547-3ec2-4442-8616-2dc3970b99b6", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-2050038244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ddafa4837ce64bf9aec427795e5f48a2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fc48e29b-113c-4849-850c-35435eab4052", "external-id": "nsx-vlan-transportzone-958", "segmentation_id": 958, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdedebbd2-90", "ovs_interfaceid": "dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.413444] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.445097] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 764.445278] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Cleaning up deleted instances with incomplete migration {{(pid=62569) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11646}} [ 764.690496] env[62569]: DEBUG oslo_concurrency.lockutils [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Releasing lock "refresh_cache-015140aa-8336-40e2-856c-723277a48f6e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.690844] env[62569]: DEBUG nova.compute.manager [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 764.691036] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 764.692434] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dcb5acf-19dc-4025-b3c9-085add47b516 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.699951] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 764.701661] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-44c9dcbe-b848-4846-93ee-f283b13f9237 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.706430] env[62569]: DEBUG oslo_vmware.api [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 764.706430] env[62569]: value = "task-1249949" [ 764.706430] env[62569]: _type = "Task" [ 764.706430] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.714485] env[62569]: DEBUG oslo_vmware.api [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249949, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.790239] env[62569]: DEBUG oslo_concurrency.lockutils [None req-23641bfb-0496-40f4-86f8-38994a4d9447 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Acquiring lock "refresh_cache-c515d85e-bcb5-4bac-bacb-1e558f38171f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.865491] env[62569]: DEBUG nova.network.neutron [-] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.880046] env[62569]: DEBUG oslo_concurrency.lockutils [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.461s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.880651] env[62569]: DEBUG nova.compute.manager [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 764.883509] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.130s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.885006] env[62569]: INFO nova.compute.claims [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 764.889727] env[62569]: DEBUG oslo_concurrency.lockutils [req-d5d6f310-509b-48c1-afa7-adfce9e8ec50 req-1620420b-bd86-4d5c-9727-7cdb01a9e5fb service nova] Releasing lock "refresh_cache-c515d85e-bcb5-4bac-bacb-1e558f38171f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.890122] env[62569]: DEBUG oslo_concurrency.lockutils [None req-23641bfb-0496-40f4-86f8-38994a4d9447 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Acquired lock "refresh_cache-c515d85e-bcb5-4bac-bacb-1e558f38171f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.890447] env[62569]: DEBUG nova.network.neutron [None req-23641bfb-0496-40f4-86f8-38994a4d9447 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 764.949080] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 765.216358] env[62569]: DEBUG oslo_vmware.api [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249949, 'name': PowerOffVM_Task, 'duration_secs': 0.196719} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.216461] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 765.216601] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 765.216853] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fe2f5c28-a0c3-44f8-8dda-728c199b5082 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.249576] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 765.249686] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 765.249810] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Deleting the datastore file [datastore1] 015140aa-8336-40e2-856c-723277a48f6e {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 765.250068] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ecc36bef-7b35-4987-aa4a-9a5d2f41079a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.256314] env[62569]: DEBUG oslo_vmware.api [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 765.256314] env[62569]: value = "task-1249951" [ 765.256314] env[62569]: _type = "Task" [ 765.256314] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.263613] env[62569]: DEBUG oslo_vmware.api [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249951, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.368448] env[62569]: INFO nova.compute.manager [-] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Took 1.03 seconds to deallocate network for instance. [ 765.392229] env[62569]: DEBUG nova.compute.utils [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 765.393627] env[62569]: DEBUG nova.compute.manager [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 765.393792] env[62569]: DEBUG nova.network.neutron [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 765.474116] env[62569]: DEBUG nova.policy [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bb83026113f247228f3c15a4e9215ee4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '548d3ffbc4ac4378a05a8adf0d0a43e7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 765.715146] env[62569]: DEBUG nova.network.neutron [None req-23641bfb-0496-40f4-86f8-38994a4d9447 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Updating instance_info_cache with network_info: [{"id": "dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897", "address": "fa:16:3e:6d:4d:fa", "network": {"id": "6135c547-3ec2-4442-8616-2dc3970b99b6", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-2050038244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ddafa4837ce64bf9aec427795e5f48a2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fc48e29b-113c-4849-850c-35435eab4052", "external-id": "nsx-vlan-transportzone-958", "segmentation_id": 958, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdedebbd2-90", "ovs_interfaceid": "dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.767564] env[62569]: DEBUG oslo_vmware.api [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1249951, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.104887} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.767564] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 765.767728] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 765.767872] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 765.768096] env[62569]: INFO nova.compute.manager [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Took 1.08 seconds to destroy the instance on the hypervisor. [ 765.768465] env[62569]: DEBUG oslo.service.loopingcall [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 765.768699] env[62569]: DEBUG nova.compute.manager [-] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 765.768803] env[62569]: DEBUG nova.network.neutron [-] [instance: 015140aa-8336-40e2-856c-723277a48f6e] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 765.786588] env[62569]: DEBUG nova.network.neutron [-] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.876013] env[62569]: DEBUG oslo_concurrency.lockutils [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.884813] env[62569]: DEBUG nova.network.neutron [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Successfully created port: 6a7245e5-dc62-4799-b00a-34cf50cd5a74 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 765.899343] env[62569]: DEBUG nova.compute.manager [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 766.220298] env[62569]: DEBUG oslo_concurrency.lockutils [None req-23641bfb-0496-40f4-86f8-38994a4d9447 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Releasing lock "refresh_cache-c515d85e-bcb5-4bac-bacb-1e558f38171f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.247960] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b96640f-11d9-4bfe-b737-c5cc5fc712cd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.256649] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f69db1-c5c0-4710-b756-5e7da4443454 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.289621] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8168c993-e39c-42cd-99e0-542629121014 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.292806] env[62569]: DEBUG nova.network.neutron [-] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.299053] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57c2413f-8abd-4b6e-affd-996b810e8c48 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.313872] env[62569]: DEBUG nova.compute.provider_tree [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.725392] env[62569]: DEBUG nova.compute.manager [None req-23641bfb-0496-40f4-86f8-38994a4d9447 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 766.726285] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d829173-5fd0-45e8-8826-2ee95a135c12 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.795384] env[62569]: INFO nova.compute.manager [-] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Took 1.03 seconds to deallocate network for instance. [ 766.816893] env[62569]: DEBUG nova.scheduler.client.report [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 766.910478] env[62569]: DEBUG nova.compute.manager [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 766.934370] env[62569]: DEBUG nova.virt.hardware [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 766.934612] env[62569]: DEBUG nova.virt.hardware [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 766.934769] env[62569]: DEBUG nova.virt.hardware [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 766.934948] env[62569]: DEBUG nova.virt.hardware [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 766.935110] env[62569]: DEBUG nova.virt.hardware [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 766.935261] env[62569]: DEBUG nova.virt.hardware [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 766.935542] env[62569]: DEBUG nova.virt.hardware [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 766.935719] env[62569]: DEBUG nova.virt.hardware [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 766.935891] env[62569]: DEBUG nova.virt.hardware [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 766.936064] env[62569]: DEBUG nova.virt.hardware [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 766.936241] env[62569]: DEBUG nova.virt.hardware [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 766.937098] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-979c926c-7aa0-4128-89b2-b2ef6e89a68b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.945297] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b33a34cf-71c7-4e97-93c7-740cb0334111 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.301526] env[62569]: DEBUG oslo_concurrency.lockutils [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.321960] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.438s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.329186] env[62569]: DEBUG nova.compute.manager [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 767.330421] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.676s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.330748] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.332942] env[62569]: DEBUG oslo_concurrency.lockutils [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.796s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.334965] env[62569]: INFO nova.compute.claims [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 767.366143] env[62569]: INFO nova.scheduler.client.report [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Deleted allocations for instance 93c8445d-689c-4f77-836f-95eafafbb6a2 [ 767.433124] env[62569]: DEBUG nova.compute.manager [req-a2065f01-ad4e-4cbf-a839-70fec505bc08 req-4c0be079-ad04-43ba-af0b-2c2444f9097d service nova] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Received event network-vif-plugged-6a7245e5-dc62-4799-b00a-34cf50cd5a74 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 767.433370] env[62569]: DEBUG oslo_concurrency.lockutils [req-a2065f01-ad4e-4cbf-a839-70fec505bc08 req-4c0be079-ad04-43ba-af0b-2c2444f9097d service nova] Acquiring lock "77a1b192-6aff-4fee-93d7-57cebcdce626-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.433617] env[62569]: DEBUG oslo_concurrency.lockutils [req-a2065f01-ad4e-4cbf-a839-70fec505bc08 req-4c0be079-ad04-43ba-af0b-2c2444f9097d service nova] Lock "77a1b192-6aff-4fee-93d7-57cebcdce626-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.433757] env[62569]: DEBUG oslo_concurrency.lockutils [req-a2065f01-ad4e-4cbf-a839-70fec505bc08 req-4c0be079-ad04-43ba-af0b-2c2444f9097d service nova] Lock "77a1b192-6aff-4fee-93d7-57cebcdce626-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.433930] env[62569]: DEBUG nova.compute.manager [req-a2065f01-ad4e-4cbf-a839-70fec505bc08 req-4c0be079-ad04-43ba-af0b-2c2444f9097d service nova] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] No waiting events found dispatching network-vif-plugged-6a7245e5-dc62-4799-b00a-34cf50cd5a74 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 767.434131] env[62569]: WARNING nova.compute.manager [req-a2065f01-ad4e-4cbf-a839-70fec505bc08 req-4c0be079-ad04-43ba-af0b-2c2444f9097d service nova] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Received unexpected event network-vif-plugged-6a7245e5-dc62-4799-b00a-34cf50cd5a74 for instance with vm_state building and task_state spawning. [ 767.479412] env[62569]: DEBUG nova.network.neutron [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Successfully updated port: 6a7245e5-dc62-4799-b00a-34cf50cd5a74 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 767.742094] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3754297-23c7-48d7-87f8-f5766e88ccc3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.750046] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-23641bfb-0496-40f4-86f8-38994a4d9447 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Doing hard reboot of VM {{(pid=62569) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 767.750151] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-7eff3c7b-ca17-4585-89c6-d49b8053b587 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.755899] env[62569]: DEBUG oslo_vmware.api [None req-23641bfb-0496-40f4-86f8-38994a4d9447 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Waiting for the task: (returnval){ [ 767.755899] env[62569]: value = "task-1249952" [ 767.755899] env[62569]: _type = "Task" [ 767.755899] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.764226] env[62569]: DEBUG oslo_vmware.api [None req-23641bfb-0496-40f4-86f8-38994a4d9447 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': task-1249952, 'name': ResetVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.842050] env[62569]: DEBUG nova.compute.utils [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 767.845020] env[62569]: DEBUG nova.compute.manager [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 767.845020] env[62569]: DEBUG nova.network.neutron [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 767.874413] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7207d9c1-7d47-464f-bb65-dbfcd55a0de0 tempest-ServerAddressesTestJSON-897153727 tempest-ServerAddressesTestJSON-897153727-project-member] Lock "93c8445d-689c-4f77-836f-95eafafbb6a2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.657s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.886213] env[62569]: DEBUG nova.policy [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a272992a4894805bfb958680a37da8e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b20340a1ce0447cae1bfd8b7c28928f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 767.983938] env[62569]: DEBUG oslo_concurrency.lockutils [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Acquiring lock "refresh_cache-77a1b192-6aff-4fee-93d7-57cebcdce626" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.983938] env[62569]: DEBUG oslo_concurrency.lockutils [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Acquired lock "refresh_cache-77a1b192-6aff-4fee-93d7-57cebcdce626" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.983938] env[62569]: DEBUG nova.network.neutron [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 768.144129] env[62569]: DEBUG nova.network.neutron [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Successfully created port: e1705802-dcd4-48a2-bad5-ffbd36806564 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 768.266944] env[62569]: DEBUG oslo_vmware.api [None req-23641bfb-0496-40f4-86f8-38994a4d9447 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': task-1249952, 'name': ResetVM_Task, 'duration_secs': 0.09829} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.267711] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-23641bfb-0496-40f4-86f8-38994a4d9447 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Did hard reboot of VM {{(pid=62569) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 768.268126] env[62569]: DEBUG nova.compute.manager [None req-23641bfb-0496-40f4-86f8-38994a4d9447 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 768.269204] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7db77b1c-3c77-4787-8ece-0c33e7553e7b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.350147] env[62569]: DEBUG nova.compute.manager [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 768.530532] env[62569]: DEBUG nova.network.neutron [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.696178] env[62569]: DEBUG nova.network.neutron [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Updating instance_info_cache with network_info: [{"id": "6a7245e5-dc62-4799-b00a-34cf50cd5a74", "address": "fa:16:3e:6f:52:54", "network": {"id": "63365435-28cb-4c92-a35f-dad6842904f6", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-620879956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "548d3ffbc4ac4378a05a8adf0d0a43e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4aa1eda7-48b9-4fa2-af0b-94c718313af2", "external-id": "nsx-vlan-transportzone-502", "segmentation_id": 502, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a7245e5-dc", "ovs_interfaceid": "6a7245e5-dc62-4799-b00a-34cf50cd5a74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.755153] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a5a9c43-f844-400e-8e67-0a22b80acd81 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.762748] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25dcb254-2152-468e-876d-ef0806a2fa38 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.796215] env[62569]: DEBUG oslo_concurrency.lockutils [None req-23641bfb-0496-40f4-86f8-38994a4d9447 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Lock "c515d85e-bcb5-4bac-bacb-1e558f38171f" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.530s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.797609] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85cbf3aa-5153-469a-a900-9e66a3dda2b7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.805052] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de729c22-b94f-4eeb-be06-f4979977fcda {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.818530] env[62569]: DEBUG nova.compute.provider_tree [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 769.199314] env[62569]: DEBUG oslo_concurrency.lockutils [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Releasing lock "refresh_cache-77a1b192-6aff-4fee-93d7-57cebcdce626" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.199678] env[62569]: DEBUG nova.compute.manager [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Instance network_info: |[{"id": "6a7245e5-dc62-4799-b00a-34cf50cd5a74", "address": "fa:16:3e:6f:52:54", "network": {"id": "63365435-28cb-4c92-a35f-dad6842904f6", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-620879956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "548d3ffbc4ac4378a05a8adf0d0a43e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4aa1eda7-48b9-4fa2-af0b-94c718313af2", "external-id": "nsx-vlan-transportzone-502", "segmentation_id": 502, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a7245e5-dc", "ovs_interfaceid": "6a7245e5-dc62-4799-b00a-34cf50cd5a74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 769.200138] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6f:52:54', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4aa1eda7-48b9-4fa2-af0b-94c718313af2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6a7245e5-dc62-4799-b00a-34cf50cd5a74', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 769.207770] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Creating folder: Project (548d3ffbc4ac4378a05a8adf0d0a43e7). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 769.208164] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-508deb80-6da5-4bbe-9f60-e6cfa89ca5cb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.219356] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Created folder: Project (548d3ffbc4ac4378a05a8adf0d0a43e7) in parent group-v269330. [ 769.219545] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Creating folder: Instances. Parent ref: group-v269371. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 769.219771] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-219388b5-99fe-4723-9512-7d4e789797e1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.229466] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Created folder: Instances in parent group-v269371. [ 769.229709] env[62569]: DEBUG oslo.service.loopingcall [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 769.229904] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 769.230201] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9993b4ee-f224-4b9b-82c4-92518169aca9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.255338] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 769.255338] env[62569]: value = "task-1249955" [ 769.255338] env[62569]: _type = "Task" [ 769.255338] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.263251] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249955, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.323321] env[62569]: DEBUG nova.scheduler.client.report [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 769.359283] env[62569]: DEBUG nova.compute.manager [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 769.388948] env[62569]: DEBUG nova.virt.hardware [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 769.389418] env[62569]: DEBUG nova.virt.hardware [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 769.389594] env[62569]: DEBUG nova.virt.hardware [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 769.389864] env[62569]: DEBUG nova.virt.hardware [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 769.390275] env[62569]: DEBUG nova.virt.hardware [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 769.390739] env[62569]: DEBUG nova.virt.hardware [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 769.391046] env[62569]: DEBUG nova.virt.hardware [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 769.391285] env[62569]: DEBUG nova.virt.hardware [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 769.391629] env[62569]: DEBUG nova.virt.hardware [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 769.391870] env[62569]: DEBUG nova.virt.hardware [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 769.392121] env[62569]: DEBUG nova.virt.hardware [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 769.393062] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73033c9f-3c24-4b06-bf52-876363e907af {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.401910] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f5b09a2-5b9b-4b1d-a6cf-56034190e4a1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.646838] env[62569]: DEBUG nova.network.neutron [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Successfully updated port: e1705802-dcd4-48a2-bad5-ffbd36806564 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 769.770743] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249955, 'name': CreateVM_Task, 'duration_secs': 0.316248} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.770822] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 769.771497] env[62569]: DEBUG oslo_concurrency.lockutils [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.771647] env[62569]: DEBUG oslo_concurrency.lockutils [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.771973] env[62569]: DEBUG oslo_concurrency.lockutils [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 769.775036] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-824588e1-8601-48cc-9727-7549a3289d6a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.777315] env[62569]: DEBUG nova.compute.manager [req-db431251-d35f-478c-ab56-75ab1c60e829 req-c0bba250-a572-4a65-a44a-230b2e84d884 service nova] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Received event network-changed-6a7245e5-dc62-4799-b00a-34cf50cd5a74 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 769.777662] env[62569]: DEBUG nova.compute.manager [req-db431251-d35f-478c-ab56-75ab1c60e829 req-c0bba250-a572-4a65-a44a-230b2e84d884 service nova] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Refreshing instance network info cache due to event network-changed-6a7245e5-dc62-4799-b00a-34cf50cd5a74. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 769.777921] env[62569]: DEBUG oslo_concurrency.lockutils [req-db431251-d35f-478c-ab56-75ab1c60e829 req-c0bba250-a572-4a65-a44a-230b2e84d884 service nova] Acquiring lock "refresh_cache-77a1b192-6aff-4fee-93d7-57cebcdce626" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.777921] env[62569]: DEBUG oslo_concurrency.lockutils [req-db431251-d35f-478c-ab56-75ab1c60e829 req-c0bba250-a572-4a65-a44a-230b2e84d884 service nova] Acquired lock "refresh_cache-77a1b192-6aff-4fee-93d7-57cebcdce626" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.778015] env[62569]: DEBUG nova.network.neutron [req-db431251-d35f-478c-ab56-75ab1c60e829 req-c0bba250-a572-4a65-a44a-230b2e84d884 service nova] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Refreshing network info cache for port 6a7245e5-dc62-4799-b00a-34cf50cd5a74 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 769.784442] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Waiting for the task: (returnval){ [ 769.784442] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ad361e-064f-e473-ddb3-32aa722312fd" [ 769.784442] env[62569]: _type = "Task" [ 769.784442] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.802264] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ad361e-064f-e473-ddb3-32aa722312fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.802604] env[62569]: DEBUG nova.compute.manager [req-8267ff92-c22f-47e7-b9e0-737b0efb854e req-8e596926-a8a0-4533-a4ba-170b62698387 service nova] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Received event network-changed-dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 769.802849] env[62569]: DEBUG nova.compute.manager [req-8267ff92-c22f-47e7-b9e0-737b0efb854e req-8e596926-a8a0-4533-a4ba-170b62698387 service nova] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Refreshing instance network info cache due to event network-changed-dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 769.803077] env[62569]: DEBUG oslo_concurrency.lockutils [req-8267ff92-c22f-47e7-b9e0-737b0efb854e req-8e596926-a8a0-4533-a4ba-170b62698387 service nova] Acquiring lock "refresh_cache-c515d85e-bcb5-4bac-bacb-1e558f38171f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.803229] env[62569]: DEBUG oslo_concurrency.lockutils [req-8267ff92-c22f-47e7-b9e0-737b0efb854e req-8e596926-a8a0-4533-a4ba-170b62698387 service nova] Acquired lock "refresh_cache-c515d85e-bcb5-4bac-bacb-1e558f38171f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.803409] env[62569]: DEBUG nova.network.neutron [req-8267ff92-c22f-47e7-b9e0-737b0efb854e req-8e596926-a8a0-4533-a4ba-170b62698387 service nova] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Refreshing network info cache for port dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 769.832330] env[62569]: DEBUG oslo_concurrency.lockutils [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.499s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.832800] env[62569]: DEBUG nova.compute.manager [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 769.836231] env[62569]: DEBUG oslo_concurrency.lockutils [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.753s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.839039] env[62569]: INFO nova.compute.claims [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 770.148653] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "refresh_cache-22094c32-5f50-4f86-a77b-cd4adcf8998a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.148653] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquired lock "refresh_cache-22094c32-5f50-4f86-a77b-cd4adcf8998a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.148812] env[62569]: DEBUG nova.network.neutron [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 770.297058] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ad361e-064f-e473-ddb3-32aa722312fd, 'name': SearchDatastore_Task, 'duration_secs': 0.012809} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.297058] env[62569]: DEBUG oslo_concurrency.lockutils [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.297058] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 770.297058] env[62569]: DEBUG oslo_concurrency.lockutils [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.297467] env[62569]: DEBUG oslo_concurrency.lockutils [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.297467] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 770.297467] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2abf5621-a373-4c59-b2a2-110c393bbf08 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.308800] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 770.308800] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 770.309772] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc7d0146-6cbd-4270-a3d1-7c0f1b341e11 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.315719] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Waiting for the task: (returnval){ [ 770.315719] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52505af3-3a0c-00ed-2faa-2d583bbda9fa" [ 770.315719] env[62569]: _type = "Task" [ 770.315719] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.326203] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52505af3-3a0c-00ed-2faa-2d583bbda9fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.350190] env[62569]: DEBUG nova.compute.utils [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 770.351952] env[62569]: DEBUG nova.compute.manager [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 770.352858] env[62569]: DEBUG nova.network.neutron [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 770.414583] env[62569]: DEBUG nova.policy [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9a84a84eb60e4fb8a49cd941af26eabd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '16ea6572453f47439342f53178c353a6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 770.615768] env[62569]: DEBUG nova.network.neutron [req-db431251-d35f-478c-ab56-75ab1c60e829 req-c0bba250-a572-4a65-a44a-230b2e84d884 service nova] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Updated VIF entry in instance network info cache for port 6a7245e5-dc62-4799-b00a-34cf50cd5a74. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 770.616142] env[62569]: DEBUG nova.network.neutron [req-db431251-d35f-478c-ab56-75ab1c60e829 req-c0bba250-a572-4a65-a44a-230b2e84d884 service nova] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Updating instance_info_cache with network_info: [{"id": "6a7245e5-dc62-4799-b00a-34cf50cd5a74", "address": "fa:16:3e:6f:52:54", "network": {"id": "63365435-28cb-4c92-a35f-dad6842904f6", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-620879956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "548d3ffbc4ac4378a05a8adf0d0a43e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4aa1eda7-48b9-4fa2-af0b-94c718313af2", "external-id": "nsx-vlan-transportzone-502", "segmentation_id": 502, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a7245e5-dc", "ovs_interfaceid": "6a7245e5-dc62-4799-b00a-34cf50cd5a74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.630732] env[62569]: DEBUG oslo_concurrency.lockutils [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Acquiring lock "c515d85e-bcb5-4bac-bacb-1e558f38171f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.632069] env[62569]: DEBUG oslo_concurrency.lockutils [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Lock "c515d85e-bcb5-4bac-bacb-1e558f38171f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.632300] env[62569]: DEBUG oslo_concurrency.lockutils [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Acquiring lock "c515d85e-bcb5-4bac-bacb-1e558f38171f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.633029] env[62569]: DEBUG oslo_concurrency.lockutils [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Lock "c515d85e-bcb5-4bac-bacb-1e558f38171f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.633241] env[62569]: DEBUG oslo_concurrency.lockutils [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Lock "c515d85e-bcb5-4bac-bacb-1e558f38171f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.635324] env[62569]: INFO nova.compute.manager [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Terminating instance [ 770.648947] env[62569]: DEBUG nova.network.neutron [req-8267ff92-c22f-47e7-b9e0-737b0efb854e req-8e596926-a8a0-4533-a4ba-170b62698387 service nova] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Updated VIF entry in instance network info cache for port dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 770.649416] env[62569]: DEBUG nova.network.neutron [req-8267ff92-c22f-47e7-b9e0-737b0efb854e req-8e596926-a8a0-4533-a4ba-170b62698387 service nova] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Updating instance_info_cache with network_info: [{"id": "dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897", "address": "fa:16:3e:6d:4d:fa", "network": {"id": "6135c547-3ec2-4442-8616-2dc3970b99b6", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-2050038244-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ddafa4837ce64bf9aec427795e5f48a2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fc48e29b-113c-4849-850c-35435eab4052", "external-id": "nsx-vlan-transportzone-958", "segmentation_id": 958, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdedebbd2-90", "ovs_interfaceid": "dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.683452] env[62569]: DEBUG nova.network.neutron [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.828015] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52505af3-3a0c-00ed-2faa-2d583bbda9fa, 'name': SearchDatastore_Task, 'duration_secs': 0.009483} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.829389] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d30e4661-1566-4165-87ca-541dee9096a9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.836195] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Waiting for the task: (returnval){ [ 770.836195] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5262f56a-87e8-ff04-a301-be1aca94f9e5" [ 770.836195] env[62569]: _type = "Task" [ 770.836195] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.841439] env[62569]: DEBUG nova.network.neutron [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Updating instance_info_cache with network_info: [{"id": "e1705802-dcd4-48a2-bad5-ffbd36806564", "address": "fa:16:3e:e4:a2:c5", "network": {"id": "8ea5e138-813b-4c5c-88a1-b1ea79807f10", "bridge": "br-int", "label": "tempest-ImagesTestJSON-524163445-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b20340a1ce0447cae1bfd8b7c28928f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1705802-dc", "ovs_interfaceid": "e1705802-dcd4-48a2-bad5-ffbd36806564", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.842929] env[62569]: DEBUG nova.network.neutron [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Successfully created port: ac6451ef-1971-4adb-90a6-9de4848d5c8d {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 770.855318] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5262f56a-87e8-ff04-a301-be1aca94f9e5, 'name': SearchDatastore_Task, 'duration_secs': 0.01038} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.856315] env[62569]: DEBUG nova.compute.manager [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 770.858735] env[62569]: DEBUG oslo_concurrency.lockutils [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.858735] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 77a1b192-6aff-4fee-93d7-57cebcdce626/77a1b192-6aff-4fee-93d7-57cebcdce626.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 770.859559] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3844c16c-5e04-4ec9-bab2-4c07f0100da0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.867097] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Waiting for the task: (returnval){ [ 770.867097] env[62569]: value = "task-1249956" [ 770.867097] env[62569]: _type = "Task" [ 770.867097] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.876992] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Task: {'id': task-1249956, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.118941] env[62569]: DEBUG oslo_concurrency.lockutils [req-db431251-d35f-478c-ab56-75ab1c60e829 req-c0bba250-a572-4a65-a44a-230b2e84d884 service nova] Releasing lock "refresh_cache-77a1b192-6aff-4fee-93d7-57cebcdce626" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.143054] env[62569]: DEBUG nova.compute.manager [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 771.143288] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 771.144206] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18c9fa99-e648-430d-ae98-eec70500198e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.151888] env[62569]: DEBUG oslo_concurrency.lockutils [req-8267ff92-c22f-47e7-b9e0-737b0efb854e req-8e596926-a8a0-4533-a4ba-170b62698387 service nova] Releasing lock "refresh_cache-c515d85e-bcb5-4bac-bacb-1e558f38171f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.155816] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 771.156550] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-436c9bfe-5d7e-4018-ad4b-16f593004759 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.166733] env[62569]: DEBUG oslo_vmware.api [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Waiting for the task: (returnval){ [ 771.166733] env[62569]: value = "task-1249957" [ 771.166733] env[62569]: _type = "Task" [ 771.166733] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.177000] env[62569]: DEBUG oslo_vmware.api [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': task-1249957, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.213324] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a7e2fe-49f4-43bf-af65-3ec096ebf032 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.225720] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e365660-a124-45c8-9fbc-aeb657338944 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.262486] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6236bdec-5571-4028-9af4-f8426d06e2e0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.271144] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7be3094-82f1-4396-b3f4-194d5a76b02d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.286464] env[62569]: DEBUG nova.compute.provider_tree [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 771.345847] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Releasing lock "refresh_cache-22094c32-5f50-4f86-a77b-cd4adcf8998a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.346211] env[62569]: DEBUG nova.compute.manager [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Instance network_info: |[{"id": "e1705802-dcd4-48a2-bad5-ffbd36806564", "address": "fa:16:3e:e4:a2:c5", "network": {"id": "8ea5e138-813b-4c5c-88a1-b1ea79807f10", "bridge": "br-int", "label": "tempest-ImagesTestJSON-524163445-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b20340a1ce0447cae1bfd8b7c28928f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1705802-dc", "ovs_interfaceid": "e1705802-dcd4-48a2-bad5-ffbd36806564", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 771.346633] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e4:a2:c5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f65996a3-f865-4492-9377-cd14ec8b3aae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e1705802-dcd4-48a2-bad5-ffbd36806564', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 771.354559] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Creating folder: Project (7b20340a1ce0447cae1bfd8b7c28928f). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 771.354856] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c8a3aa9e-a376-4646-ae72-ef5d80e9ede9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.367288] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Created folder: Project (7b20340a1ce0447cae1bfd8b7c28928f) in parent group-v269330. [ 771.367288] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Creating folder: Instances. Parent ref: group-v269374. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 771.367288] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-46596c11-6c14-4ad0-9e93-05afda3eb251 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.378614] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Task: {'id': task-1249956, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45423} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.382059] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 77a1b192-6aff-4fee-93d7-57cebcdce626/77a1b192-6aff-4fee-93d7-57cebcdce626.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 771.382059] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 771.382059] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Created folder: Instances in parent group-v269374. [ 771.382059] env[62569]: DEBUG oslo.service.loopingcall [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.382059] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b34c42ec-31bf-4968-87c9-0785a0723e32 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.382867] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 771.383073] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fb05b587-ee26-4d0a-a4a9-9cf36f3328fc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.402831] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 771.402831] env[62569]: value = "task-1249961" [ 771.402831] env[62569]: _type = "Task" [ 771.402831] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.404153] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Waiting for the task: (returnval){ [ 771.404153] env[62569]: value = "task-1249960" [ 771.404153] env[62569]: _type = "Task" [ 771.404153] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.414438] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249961, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.417764] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Task: {'id': task-1249960, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.678028] env[62569]: DEBUG oslo_vmware.api [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': task-1249957, 'name': PowerOffVM_Task, 'duration_secs': 0.230727} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.678323] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 771.678526] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 771.678777] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3de8c002-69e8-417d-aaa2-0cc91b8d39d2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.768996] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 771.769229] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 771.769416] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Deleting the datastore file [datastore2] c515d85e-bcb5-4bac-bacb-1e558f38171f {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 771.769792] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f8606e41-d8fd-4237-9b73-d5e9aa6abb53 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.777493] env[62569]: DEBUG oslo_vmware.api [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Waiting for the task: (returnval){ [ 771.777493] env[62569]: value = "task-1249963" [ 771.777493] env[62569]: _type = "Task" [ 771.777493] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.785666] env[62569]: DEBUG oslo_vmware.api [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': task-1249963, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.789633] env[62569]: DEBUG nova.scheduler.client.report [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 771.866173] env[62569]: DEBUG nova.compute.manager [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 771.910864] env[62569]: DEBUG nova.virt.hardware [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 771.911161] env[62569]: DEBUG nova.virt.hardware [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 771.911328] env[62569]: DEBUG nova.virt.hardware [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 771.911516] env[62569]: DEBUG nova.virt.hardware [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 771.911672] env[62569]: DEBUG nova.virt.hardware [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 771.912010] env[62569]: DEBUG nova.virt.hardware [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 771.912199] env[62569]: DEBUG nova.virt.hardware [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 771.912352] env[62569]: DEBUG nova.virt.hardware [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 771.912569] env[62569]: DEBUG nova.virt.hardware [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 771.912746] env[62569]: DEBUG nova.virt.hardware [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 771.912922] env[62569]: DEBUG nova.virt.hardware [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 771.913903] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bde52a3-3f64-4ef7-b0c2-5a257ac8dfbc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.929939] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75784326-40cd-4e73-bc1d-c17bd8292101 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.933899] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249961, 'name': CreateVM_Task, 'duration_secs': 0.315746} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.934151] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Task: {'id': task-1249960, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06417} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.934627] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 771.934960] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 771.935931] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.936061] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.936367] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 771.937165] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad37d70a-3892-478a-8af7-bc87f9ff19b1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.947382] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63e9f117-8847-41be-a9fa-1b398ba5dda2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.968970] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] 77a1b192-6aff-4fee-93d7-57cebcdce626/77a1b192-6aff-4fee-93d7-57cebcdce626.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 771.969971] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a382f71-eb8a-4add-a4c7-3cf5bd189351 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.986299] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 771.986299] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5225dfbc-adc4-0294-a2ba-5d0c96fe98d8" [ 771.986299] env[62569]: _type = "Task" [ 771.986299] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.989197] env[62569]: DEBUG nova.compute.manager [req-10331d8c-ede5-42ac-b17c-7eaf7e1abf4d req-d2780ba6-bd23-4077-ac2e-ba06afcb6b94 service nova] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Received event network-vif-plugged-e1705802-dcd4-48a2-bad5-ffbd36806564 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 771.989197] env[62569]: DEBUG oslo_concurrency.lockutils [req-10331d8c-ede5-42ac-b17c-7eaf7e1abf4d req-d2780ba6-bd23-4077-ac2e-ba06afcb6b94 service nova] Acquiring lock "22094c32-5f50-4f86-a77b-cd4adcf8998a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.989197] env[62569]: DEBUG oslo_concurrency.lockutils [req-10331d8c-ede5-42ac-b17c-7eaf7e1abf4d req-d2780ba6-bd23-4077-ac2e-ba06afcb6b94 service nova] Lock "22094c32-5f50-4f86-a77b-cd4adcf8998a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.989197] env[62569]: DEBUG oslo_concurrency.lockutils [req-10331d8c-ede5-42ac-b17c-7eaf7e1abf4d req-d2780ba6-bd23-4077-ac2e-ba06afcb6b94 service nova] Lock "22094c32-5f50-4f86-a77b-cd4adcf8998a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.989374] env[62569]: DEBUG nova.compute.manager [req-10331d8c-ede5-42ac-b17c-7eaf7e1abf4d req-d2780ba6-bd23-4077-ac2e-ba06afcb6b94 service nova] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] No waiting events found dispatching network-vif-plugged-e1705802-dcd4-48a2-bad5-ffbd36806564 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 771.989406] env[62569]: WARNING nova.compute.manager [req-10331d8c-ede5-42ac-b17c-7eaf7e1abf4d req-d2780ba6-bd23-4077-ac2e-ba06afcb6b94 service nova] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Received unexpected event network-vif-plugged-e1705802-dcd4-48a2-bad5-ffbd36806564 for instance with vm_state building and task_state spawning. [ 771.989555] env[62569]: DEBUG nova.compute.manager [req-10331d8c-ede5-42ac-b17c-7eaf7e1abf4d req-d2780ba6-bd23-4077-ac2e-ba06afcb6b94 service nova] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Received event network-changed-e1705802-dcd4-48a2-bad5-ffbd36806564 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 771.989712] env[62569]: DEBUG nova.compute.manager [req-10331d8c-ede5-42ac-b17c-7eaf7e1abf4d req-d2780ba6-bd23-4077-ac2e-ba06afcb6b94 service nova] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Refreshing instance network info cache due to event network-changed-e1705802-dcd4-48a2-bad5-ffbd36806564. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 771.989882] env[62569]: DEBUG oslo_concurrency.lockutils [req-10331d8c-ede5-42ac-b17c-7eaf7e1abf4d req-d2780ba6-bd23-4077-ac2e-ba06afcb6b94 service nova] Acquiring lock "refresh_cache-22094c32-5f50-4f86-a77b-cd4adcf8998a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.990027] env[62569]: DEBUG oslo_concurrency.lockutils [req-10331d8c-ede5-42ac-b17c-7eaf7e1abf4d req-d2780ba6-bd23-4077-ac2e-ba06afcb6b94 service nova] Acquired lock "refresh_cache-22094c32-5f50-4f86-a77b-cd4adcf8998a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.990175] env[62569]: DEBUG nova.network.neutron [req-10331d8c-ede5-42ac-b17c-7eaf7e1abf4d req-d2780ba6-bd23-4077-ac2e-ba06afcb6b94 service nova] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Refreshing network info cache for port e1705802-dcd4-48a2-bad5-ffbd36806564 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 771.997192] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Waiting for the task: (returnval){ [ 771.997192] env[62569]: value = "task-1249964" [ 771.997192] env[62569]: _type = "Task" [ 771.997192] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.002297] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5225dfbc-adc4-0294-a2ba-5d0c96fe98d8, 'name': SearchDatastore_Task, 'duration_secs': 0.009237} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.007282] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.007541] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 772.007862] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.008116] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.008220] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 772.008873] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7500bb95-2487-488c-9eac-4c4e2a3f3135 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.017074] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Task: {'id': task-1249964, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.018839] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 772.019058] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 772.019867] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91da701d-00c2-4548-817a-18c0fdfa7d03 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.025292] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 772.025292] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528397aa-5d3d-4969-a8f6-df8769089a82" [ 772.025292] env[62569]: _type = "Task" [ 772.025292] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.032869] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528397aa-5d3d-4969-a8f6-df8769089a82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.287370] env[62569]: DEBUG oslo_vmware.api [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Task: {'id': task-1249963, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145765} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.287726] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 772.287991] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 772.288276] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 772.288523] env[62569]: INFO nova.compute.manager [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Took 1.15 seconds to destroy the instance on the hypervisor. [ 772.288844] env[62569]: DEBUG oslo.service.loopingcall [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 772.289126] env[62569]: DEBUG nova.compute.manager [-] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 772.289281] env[62569]: DEBUG nova.network.neutron [-] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 772.294022] env[62569]: DEBUG oslo_concurrency.lockutils [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.458s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.294612] env[62569]: DEBUG nova.compute.manager [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 772.298553] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.713s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.300469] env[62569]: INFO nova.compute.claims [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 772.510672] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Task: {'id': task-1249964, 'name': ReconfigVM_Task, 'duration_secs': 0.268273} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.511085] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Reconfigured VM instance instance-00000039 to attach disk [datastore1] 77a1b192-6aff-4fee-93d7-57cebcdce626/77a1b192-6aff-4fee-93d7-57cebcdce626.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 772.511757] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-45c978ab-4052-41a7-b934-3e6c799201cb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.518088] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Waiting for the task: (returnval){ [ 772.518088] env[62569]: value = "task-1249965" [ 772.518088] env[62569]: _type = "Task" [ 772.518088] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.526341] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Task: {'id': task-1249965, 'name': Rename_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.535244] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528397aa-5d3d-4969-a8f6-df8769089a82, 'name': SearchDatastore_Task, 'duration_secs': 0.008338} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.535997] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1bb0074-f208-423b-b9c5-7db6f430f35a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.541838] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 772.541838] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]523fa1af-30c1-cbe7-7cf4-b50b6b51f1a9" [ 772.541838] env[62569]: _type = "Task" [ 772.541838] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.551624] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]523fa1af-30c1-cbe7-7cf4-b50b6b51f1a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.805927] env[62569]: DEBUG nova.compute.utils [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 772.811802] env[62569]: DEBUG nova.compute.manager [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 772.811802] env[62569]: DEBUG nova.network.neutron [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 772.817664] env[62569]: DEBUG nova.compute.manager [req-28af4b31-31c4-402e-b67e-7c4b8a811696 req-e959910a-edd6-44c4-b8b5-5e19eeedfa85 service nova] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Received event network-vif-plugged-ac6451ef-1971-4adb-90a6-9de4848d5c8d {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 772.817861] env[62569]: DEBUG oslo_concurrency.lockutils [req-28af4b31-31c4-402e-b67e-7c4b8a811696 req-e959910a-edd6-44c4-b8b5-5e19eeedfa85 service nova] Acquiring lock "4feb7b78-9f7c-4e64-b0a7-870ed73adf97-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.818071] env[62569]: DEBUG oslo_concurrency.lockutils [req-28af4b31-31c4-402e-b67e-7c4b8a811696 req-e959910a-edd6-44c4-b8b5-5e19eeedfa85 service nova] Lock "4feb7b78-9f7c-4e64-b0a7-870ed73adf97-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.818265] env[62569]: DEBUG oslo_concurrency.lockutils [req-28af4b31-31c4-402e-b67e-7c4b8a811696 req-e959910a-edd6-44c4-b8b5-5e19eeedfa85 service nova] Lock "4feb7b78-9f7c-4e64-b0a7-870ed73adf97-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.819015] env[62569]: DEBUG nova.compute.manager [req-28af4b31-31c4-402e-b67e-7c4b8a811696 req-e959910a-edd6-44c4-b8b5-5e19eeedfa85 service nova] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] No waiting events found dispatching network-vif-plugged-ac6451ef-1971-4adb-90a6-9de4848d5c8d {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 772.819015] env[62569]: WARNING nova.compute.manager [req-28af4b31-31c4-402e-b67e-7c4b8a811696 req-e959910a-edd6-44c4-b8b5-5e19eeedfa85 service nova] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Received unexpected event network-vif-plugged-ac6451ef-1971-4adb-90a6-9de4848d5c8d for instance with vm_state building and task_state spawning. [ 772.882184] env[62569]: DEBUG nova.network.neutron [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Successfully updated port: ac6451ef-1971-4adb-90a6-9de4848d5c8d {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 772.914463] env[62569]: DEBUG nova.policy [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '899e8bb7962a499a8ae2ab768eaec37e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fe13e05a444f4f1d94ce9ca68153703e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 772.956256] env[62569]: DEBUG nova.network.neutron [req-10331d8c-ede5-42ac-b17c-7eaf7e1abf4d req-d2780ba6-bd23-4077-ac2e-ba06afcb6b94 service nova] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Updated VIF entry in instance network info cache for port e1705802-dcd4-48a2-bad5-ffbd36806564. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 772.957381] env[62569]: DEBUG nova.network.neutron [req-10331d8c-ede5-42ac-b17c-7eaf7e1abf4d req-d2780ba6-bd23-4077-ac2e-ba06afcb6b94 service nova] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Updating instance_info_cache with network_info: [{"id": "e1705802-dcd4-48a2-bad5-ffbd36806564", "address": "fa:16:3e:e4:a2:c5", "network": {"id": "8ea5e138-813b-4c5c-88a1-b1ea79807f10", "bridge": "br-int", "label": "tempest-ImagesTestJSON-524163445-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b20340a1ce0447cae1bfd8b7c28928f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1705802-dc", "ovs_interfaceid": "e1705802-dcd4-48a2-bad5-ffbd36806564", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.034691] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Task: {'id': task-1249965, 'name': Rename_Task, 'duration_secs': 0.140666} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.035449] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 773.035732] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8e880381-6d51-4627-9a2d-4b9561a6bae6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.042966] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Waiting for the task: (returnval){ [ 773.042966] env[62569]: value = "task-1249966" [ 773.042966] env[62569]: _type = "Task" [ 773.042966] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.056765] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]523fa1af-30c1-cbe7-7cf4-b50b6b51f1a9, 'name': SearchDatastore_Task, 'duration_secs': 0.009263} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.060080] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.060557] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 22094c32-5f50-4f86-a77b-cd4adcf8998a/22094c32-5f50-4f86-a77b-cd4adcf8998a.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 773.060938] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Task: {'id': task-1249966, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.061260] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0991b992-ffa2-4c53-9fa9-516013c16c1a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.068053] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 773.068053] env[62569]: value = "task-1249967" [ 773.068053] env[62569]: _type = "Task" [ 773.068053] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.078865] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1249967, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.280660] env[62569]: DEBUG nova.network.neutron [-] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.311918] env[62569]: DEBUG nova.compute.manager [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 773.391387] env[62569]: DEBUG oslo_concurrency.lockutils [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Acquiring lock "refresh_cache-4feb7b78-9f7c-4e64-b0a7-870ed73adf97" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.391656] env[62569]: DEBUG oslo_concurrency.lockutils [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Acquired lock "refresh_cache-4feb7b78-9f7c-4e64-b0a7-870ed73adf97" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.391818] env[62569]: DEBUG nova.network.neutron [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 773.460400] env[62569]: DEBUG oslo_concurrency.lockutils [req-10331d8c-ede5-42ac-b17c-7eaf7e1abf4d req-d2780ba6-bd23-4077-ac2e-ba06afcb6b94 service nova] Releasing lock "refresh_cache-22094c32-5f50-4f86-a77b-cd4adcf8998a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.558743] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Task: {'id': task-1249966, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.583777] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1249967, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.642426] env[62569]: DEBUG nova.network.neutron [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Successfully created port: 62cedd28-f07d-4ca0-89b4-ffbc290ff45f {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 773.758411] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da2f6193-e238-4c13-8ae0-90d552d8083a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.767014] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4a18ef4-6201-432c-b781-1f235590f863 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.799223] env[62569]: INFO nova.compute.manager [-] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Took 1.51 seconds to deallocate network for instance. [ 773.801961] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93086e44-6cf8-4c8e-85f4-d79e8eee53a6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.812869] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80bc537b-f17e-4ac9-a78f-d5da27be56c9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.832625] env[62569]: DEBUG nova.compute.provider_tree [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.956701] env[62569]: DEBUG nova.network.neutron [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.061291] env[62569]: DEBUG oslo_vmware.api [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Task: {'id': task-1249966, 'name': PowerOnVM_Task, 'duration_secs': 0.566229} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.062941] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 774.063175] env[62569]: INFO nova.compute.manager [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Took 7.15 seconds to spawn the instance on the hypervisor. [ 774.063361] env[62569]: DEBUG nova.compute.manager [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 774.065144] env[62569]: DEBUG nova.compute.manager [req-fe202bc3-c868-4ef0-be4a-b882976b4853 req-fea82b02-b364-498d-9b0f-491453577cdc service nova] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Received event network-vif-deleted-dedebbd2-90a0-4ba1-b4c9-a1a4ab7c7897 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 774.065391] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df79616e-bd9f-42f5-a884-192c97bf1f30 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.086041] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1249967, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.688038} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.086041] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 22094c32-5f50-4f86-a77b-cd4adcf8998a/22094c32-5f50-4f86-a77b-cd4adcf8998a.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 774.086041] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 774.086218] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e463bab8-74d1-4296-87a4-bbbb27ebf85d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.095783] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 774.095783] env[62569]: value = "task-1249968" [ 774.095783] env[62569]: _type = "Task" [ 774.095783] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.104258] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1249968, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.280089] env[62569]: DEBUG nova.network.neutron [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Updating instance_info_cache with network_info: [{"id": "ac6451ef-1971-4adb-90a6-9de4848d5c8d", "address": "fa:16:3e:de:d5:be", "network": {"id": "8722beaf-8870-4e0e-9a9a-7a0f17bacaab", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-48483054-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16ea6572453f47439342f53178c353a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff1f3320-df8e-49df-a412-9797a23bd173", "external-id": "nsx-vlan-transportzone-217", "segmentation_id": 217, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac6451ef-19", "ovs_interfaceid": "ac6451ef-1971-4adb-90a6-9de4848d5c8d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.309935] env[62569]: DEBUG oslo_concurrency.lockutils [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.336327] env[62569]: DEBUG nova.compute.manager [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 774.339805] env[62569]: DEBUG nova.scheduler.client.report [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 774.366792] env[62569]: DEBUG nova.virt.hardware [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 774.367609] env[62569]: DEBUG nova.virt.hardware [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 774.367609] env[62569]: DEBUG nova.virt.hardware [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 774.367609] env[62569]: DEBUG nova.virt.hardware [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 774.367609] env[62569]: DEBUG nova.virt.hardware [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 774.367758] env[62569]: DEBUG nova.virt.hardware [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 774.367865] env[62569]: DEBUG nova.virt.hardware [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 774.368121] env[62569]: DEBUG nova.virt.hardware [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 774.368288] env[62569]: DEBUG nova.virt.hardware [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 774.368455] env[62569]: DEBUG nova.virt.hardware [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 774.368635] env[62569]: DEBUG nova.virt.hardware [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 774.369818] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cce2de1-ffb8-4096-a086-b51adc55f645 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.378192] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef7fb33a-73c0-47c1-ac56-e80d45fa15b5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.592544] env[62569]: INFO nova.compute.manager [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Took 33.37 seconds to build instance. [ 774.606696] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1249968, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.27034} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.606979] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 774.607880] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90fda278-54f5-4b45-b983-f11b8d44c328 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.638108] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] 22094c32-5f50-4f86-a77b-cd4adcf8998a/22094c32-5f50-4f86-a77b-cd4adcf8998a.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 774.638684] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3c66d9d-96cb-4361-9066-58488c0d5f10 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.670021] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 774.670021] env[62569]: value = "task-1249969" [ 774.670021] env[62569]: _type = "Task" [ 774.670021] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.676893] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1249969, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.786522] env[62569]: DEBUG oslo_concurrency.lockutils [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Releasing lock "refresh_cache-4feb7b78-9f7c-4e64-b0a7-870ed73adf97" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.786522] env[62569]: DEBUG nova.compute.manager [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Instance network_info: |[{"id": "ac6451ef-1971-4adb-90a6-9de4848d5c8d", "address": "fa:16:3e:de:d5:be", "network": {"id": "8722beaf-8870-4e0e-9a9a-7a0f17bacaab", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-48483054-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16ea6572453f47439342f53178c353a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff1f3320-df8e-49df-a412-9797a23bd173", "external-id": "nsx-vlan-transportzone-217", "segmentation_id": 217, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac6451ef-19", "ovs_interfaceid": "ac6451ef-1971-4adb-90a6-9de4848d5c8d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 774.786643] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:de:d5:be', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ff1f3320-df8e-49df-a412-9797a23bd173', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ac6451ef-1971-4adb-90a6-9de4848d5c8d', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 774.793845] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Creating folder: Project (16ea6572453f47439342f53178c353a6). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 774.794279] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-84dc165e-103c-4ebc-ad43-802f49931e89 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.807021] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Created folder: Project (16ea6572453f47439342f53178c353a6) in parent group-v269330. [ 774.807021] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Creating folder: Instances. Parent ref: group-v269377. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 774.807021] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-169f975f-add2-46ef-a8c3-6501cd742418 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.816160] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Created folder: Instances in parent group-v269377. [ 774.818023] env[62569]: DEBUG oslo.service.loopingcall [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 774.818023] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 774.818023] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5af4ca1f-d399-486f-9c82-344281860c35 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.839061] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 774.839061] env[62569]: value = "task-1249972" [ 774.839061] env[62569]: _type = "Task" [ 774.839061] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.848847] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.550s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.849388] env[62569]: DEBUG nova.compute.manager [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 774.852325] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249972, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.852691] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.240s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.856160] env[62569]: INFO nova.compute.claims [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 774.908606] env[62569]: DEBUG nova.compute.manager [req-d36045dd-8592-4850-ac1e-82494d8bc61e req-7a150ae1-1c29-49e2-a3ea-e1f21fd12497 service nova] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Received event network-changed-ac6451ef-1971-4adb-90a6-9de4848d5c8d {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 774.908857] env[62569]: DEBUG nova.compute.manager [req-d36045dd-8592-4850-ac1e-82494d8bc61e req-7a150ae1-1c29-49e2-a3ea-e1f21fd12497 service nova] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Refreshing instance network info cache due to event network-changed-ac6451ef-1971-4adb-90a6-9de4848d5c8d. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 774.909114] env[62569]: DEBUG oslo_concurrency.lockutils [req-d36045dd-8592-4850-ac1e-82494d8bc61e req-7a150ae1-1c29-49e2-a3ea-e1f21fd12497 service nova] Acquiring lock "refresh_cache-4feb7b78-9f7c-4e64-b0a7-870ed73adf97" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.909265] env[62569]: DEBUG oslo_concurrency.lockutils [req-d36045dd-8592-4850-ac1e-82494d8bc61e req-7a150ae1-1c29-49e2-a3ea-e1f21fd12497 service nova] Acquired lock "refresh_cache-4feb7b78-9f7c-4e64-b0a7-870ed73adf97" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.910028] env[62569]: DEBUG nova.network.neutron [req-d36045dd-8592-4850-ac1e-82494d8bc61e req-7a150ae1-1c29-49e2-a3ea-e1f21fd12497 service nova] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Refreshing network info cache for port ac6451ef-1971-4adb-90a6-9de4848d5c8d {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 775.096301] env[62569]: DEBUG oslo_concurrency.lockutils [None req-db8044a9-0e34-4f6e-b63a-529dc98f45a9 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Lock "77a1b192-6aff-4fee-93d7-57cebcdce626" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.834s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.181498] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1249969, 'name': ReconfigVM_Task, 'duration_secs': 0.36549} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.181498] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Reconfigured VM instance instance-0000003a to attach disk [datastore1] 22094c32-5f50-4f86-a77b-cd4adcf8998a/22094c32-5f50-4f86-a77b-cd4adcf8998a.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 775.182051] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e79c35f5-413d-4b1c-8e05-dbaa78d6e6d0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.188809] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 775.188809] env[62569]: value = "task-1249973" [ 775.188809] env[62569]: _type = "Task" [ 775.188809] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.198631] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1249973, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.356352] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249972, 'name': CreateVM_Task, 'duration_secs': 0.321001} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.356352] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 775.356352] env[62569]: DEBUG oslo_concurrency.lockutils [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.356352] env[62569]: DEBUG oslo_concurrency.lockutils [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.356352] env[62569]: DEBUG oslo_concurrency.lockutils [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 775.356590] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5fed5994-25f8-4b18-a260-12aec24fd5a9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.359159] env[62569]: DEBUG nova.compute.utils [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 775.365894] env[62569]: DEBUG nova.compute.manager [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 775.366127] env[62569]: DEBUG nova.network.neutron [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 775.373370] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Waiting for the task: (returnval){ [ 775.373370] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5293229e-bb93-19cf-52e6-951f7d4f63b9" [ 775.373370] env[62569]: _type = "Task" [ 775.373370] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.389682] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5293229e-bb93-19cf-52e6-951f7d4f63b9, 'name': SearchDatastore_Task, 'duration_secs': 0.009627} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.390051] env[62569]: DEBUG oslo_concurrency.lockutils [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.390440] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 775.390740] env[62569]: DEBUG oslo_concurrency.lockutils [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.390891] env[62569]: DEBUG oslo_concurrency.lockutils [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.391080] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 775.391348] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-197bec34-6b6e-4f74-8d4e-4eda690e9480 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.399699] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 775.399871] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 775.400640] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f899bddd-875e-4b25-bfea-a40e1325eff5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.405832] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Waiting for the task: (returnval){ [ 775.405832] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]521b618e-a1c2-5afa-b49c-899abf2b9113" [ 775.405832] env[62569]: _type = "Task" [ 775.405832] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.416747] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]521b618e-a1c2-5afa-b49c-899abf2b9113, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.456742] env[62569]: DEBUG nova.policy [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e5307b1d2b69421ab908a0d655a593fe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'de8605118d744a93bce54e897aa849b5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 775.598730] env[62569]: DEBUG nova.compute.manager [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 775.701439] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1249973, 'name': Rename_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.863854] env[62569]: DEBUG nova.compute.manager [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 775.885191] env[62569]: DEBUG nova.network.neutron [req-d36045dd-8592-4850-ac1e-82494d8bc61e req-7a150ae1-1c29-49e2-a3ea-e1f21fd12497 service nova] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Updated VIF entry in instance network info cache for port ac6451ef-1971-4adb-90a6-9de4848d5c8d. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 775.885191] env[62569]: DEBUG nova.network.neutron [req-d36045dd-8592-4850-ac1e-82494d8bc61e req-7a150ae1-1c29-49e2-a3ea-e1f21fd12497 service nova] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Updating instance_info_cache with network_info: [{"id": "ac6451ef-1971-4adb-90a6-9de4848d5c8d", "address": "fa:16:3e:de:d5:be", "network": {"id": "8722beaf-8870-4e0e-9a9a-7a0f17bacaab", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-48483054-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "16ea6572453f47439342f53178c353a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ff1f3320-df8e-49df-a412-9797a23bd173", "external-id": "nsx-vlan-transportzone-217", "segmentation_id": 217, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac6451ef-19", "ovs_interfaceid": "ac6451ef-1971-4adb-90a6-9de4848d5c8d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.922942] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]521b618e-a1c2-5afa-b49c-899abf2b9113, 'name': SearchDatastore_Task, 'duration_secs': 0.008123} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.923805] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-968ac796-cda0-4010-ae5e-c3fa6cb0b95e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.939024] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Waiting for the task: (returnval){ [ 775.939024] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52390c94-8f27-2cc5-7e41-714d98b6c189" [ 775.939024] env[62569]: _type = "Task" [ 775.939024] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.953681] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52390c94-8f27-2cc5-7e41-714d98b6c189, 'name': SearchDatastore_Task, 'duration_secs': 0.00929} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.956023] env[62569]: DEBUG nova.network.neutron [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Successfully updated port: 62cedd28-f07d-4ca0-89b4-ffbc290ff45f {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 775.960399] env[62569]: DEBUG oslo_concurrency.lockutils [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.960700] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 4feb7b78-9f7c-4e64-b0a7-870ed73adf97/4feb7b78-9f7c-4e64-b0a7-870ed73adf97.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 775.961951] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b38d32ec-92f1-4fcc-b5fd-ad8b5d82f28c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.970733] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Waiting for the task: (returnval){ [ 775.970733] env[62569]: value = "task-1249974" [ 775.970733] env[62569]: _type = "Task" [ 775.970733] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.978595] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Task: {'id': task-1249974, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.050137] env[62569]: DEBUG nova.network.neutron [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Successfully created port: cfddbae7-9ab0-4bc6-aea1-46a0cde743fc {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 776.130913] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.207511] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1249973, 'name': Rename_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.286456] env[62569]: DEBUG nova.compute.manager [req-57b250fa-e451-4854-92da-128a2fb99283 req-4722f54f-4821-4ae7-956b-5b87ca0b280c service nova] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Received event network-vif-plugged-62cedd28-f07d-4ca0-89b4-ffbc290ff45f {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 776.286676] env[62569]: DEBUG oslo_concurrency.lockutils [req-57b250fa-e451-4854-92da-128a2fb99283 req-4722f54f-4821-4ae7-956b-5b87ca0b280c service nova] Acquiring lock "0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.286890] env[62569]: DEBUG oslo_concurrency.lockutils [req-57b250fa-e451-4854-92da-128a2fb99283 req-4722f54f-4821-4ae7-956b-5b87ca0b280c service nova] Lock "0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.289015] env[62569]: DEBUG oslo_concurrency.lockutils [req-57b250fa-e451-4854-92da-128a2fb99283 req-4722f54f-4821-4ae7-956b-5b87ca0b280c service nova] Lock "0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.289015] env[62569]: DEBUG nova.compute.manager [req-57b250fa-e451-4854-92da-128a2fb99283 req-4722f54f-4821-4ae7-956b-5b87ca0b280c service nova] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] No waiting events found dispatching network-vif-plugged-62cedd28-f07d-4ca0-89b4-ffbc290ff45f {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 776.289015] env[62569]: WARNING nova.compute.manager [req-57b250fa-e451-4854-92da-128a2fb99283 req-4722f54f-4821-4ae7-956b-5b87ca0b280c service nova] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Received unexpected event network-vif-plugged-62cedd28-f07d-4ca0-89b4-ffbc290ff45f for instance with vm_state building and task_state spawning. [ 776.289015] env[62569]: DEBUG nova.compute.manager [req-57b250fa-e451-4854-92da-128a2fb99283 req-4722f54f-4821-4ae7-956b-5b87ca0b280c service nova] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Received event network-changed-62cedd28-f07d-4ca0-89b4-ffbc290ff45f {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 776.289015] env[62569]: DEBUG nova.compute.manager [req-57b250fa-e451-4854-92da-128a2fb99283 req-4722f54f-4821-4ae7-956b-5b87ca0b280c service nova] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Refreshing instance network info cache due to event network-changed-62cedd28-f07d-4ca0-89b4-ffbc290ff45f. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 776.289305] env[62569]: DEBUG oslo_concurrency.lockutils [req-57b250fa-e451-4854-92da-128a2fb99283 req-4722f54f-4821-4ae7-956b-5b87ca0b280c service nova] Acquiring lock "refresh_cache-0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.289305] env[62569]: DEBUG oslo_concurrency.lockutils [req-57b250fa-e451-4854-92da-128a2fb99283 req-4722f54f-4821-4ae7-956b-5b87ca0b280c service nova] Acquired lock "refresh_cache-0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.289305] env[62569]: DEBUG nova.network.neutron [req-57b250fa-e451-4854-92da-128a2fb99283 req-4722f54f-4821-4ae7-956b-5b87ca0b280c service nova] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Refreshing network info cache for port 62cedd28-f07d-4ca0-89b4-ffbc290ff45f {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 776.348738] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-319a7d2d-17a6-4fee-aafc-e0f8478ad79c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.358367] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8b4b2fa-d56a-476f-9200-a881e0263237 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.399643] env[62569]: DEBUG oslo_concurrency.lockutils [req-d36045dd-8592-4850-ac1e-82494d8bc61e req-7a150ae1-1c29-49e2-a3ea-e1f21fd12497 service nova] Releasing lock "refresh_cache-4feb7b78-9f7c-4e64-b0a7-870ed73adf97" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.401244] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c2e5d4d-1d7c-409e-954d-8ad938867978 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.411764] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc4aceb6-ef18-40bd-951e-d6b021b610ef {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.427849] env[62569]: DEBUG nova.compute.provider_tree [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.462654] env[62569]: DEBUG oslo_concurrency.lockutils [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Acquiring lock "refresh_cache-0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.484734] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Task: {'id': task-1249974, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.47296} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.485495] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 4feb7b78-9f7c-4e64-b0a7-870ed73adf97/4feb7b78-9f7c-4e64-b0a7-870ed73adf97.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 776.486191] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 776.486678] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-04bf058f-dc95-4b9b-9c2c-3767644c782e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.497068] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Waiting for the task: (returnval){ [ 776.497068] env[62569]: value = "task-1249975" [ 776.497068] env[62569]: _type = "Task" [ 776.497068] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.509064] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Task: {'id': task-1249975, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.663676] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Acquiring lock "6e83fd32-215a-42e4-99e7-72bdce25555e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.663950] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Lock "6e83fd32-215a-42e4-99e7-72bdce25555e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.700472] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1249973, 'name': Rename_Task, 'duration_secs': 1.120106} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.700472] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 776.700472] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bb56990b-fdbb-46bc-b492-f4a9255840f0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.706103] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 776.706103] env[62569]: value = "task-1249976" [ 776.706103] env[62569]: _type = "Task" [ 776.706103] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.714069] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1249976, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.828788] env[62569]: DEBUG nova.network.neutron [req-57b250fa-e451-4854-92da-128a2fb99283 req-4722f54f-4821-4ae7-956b-5b87ca0b280c service nova] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.902937] env[62569]: DEBUG nova.compute.manager [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 776.920155] env[62569]: DEBUG nova.network.neutron [req-57b250fa-e451-4854-92da-128a2fb99283 req-4722f54f-4821-4ae7-956b-5b87ca0b280c service nova] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.930113] env[62569]: DEBUG nova.virt.hardware [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 776.930400] env[62569]: DEBUG nova.virt.hardware [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 776.930569] env[62569]: DEBUG nova.virt.hardware [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 776.930759] env[62569]: DEBUG nova.virt.hardware [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 776.930908] env[62569]: DEBUG nova.virt.hardware [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 776.931071] env[62569]: DEBUG nova.virt.hardware [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 776.931285] env[62569]: DEBUG nova.virt.hardware [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 776.931506] env[62569]: DEBUG nova.virt.hardware [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 776.931701] env[62569]: DEBUG nova.virt.hardware [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 776.931870] env[62569]: DEBUG nova.virt.hardware [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 776.932056] env[62569]: DEBUG nova.virt.hardware [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 776.932874] env[62569]: DEBUG nova.scheduler.client.report [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 776.936329] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3ae0d8d-cb79-4f5a-948b-fd08c4567874 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.945173] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e957c296-6331-4a9c-8c90-d72489c4e341 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.008485] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Task: {'id': task-1249975, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067743} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.008653] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 777.009445] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ca3b495-f5fb-4206-ac4f-67529560c4be {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.031815] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] 4feb7b78-9f7c-4e64-b0a7-870ed73adf97/4feb7b78-9f7c-4e64-b0a7-870ed73adf97.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 777.032183] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9407c6d3-ed23-4b1f-8a06-992f49425f2d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.053855] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Waiting for the task: (returnval){ [ 777.053855] env[62569]: value = "task-1249977" [ 777.053855] env[62569]: _type = "Task" [ 777.053855] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.061864] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Task: {'id': task-1249977, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.216284] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1249976, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.428024] env[62569]: DEBUG oslo_concurrency.lockutils [req-57b250fa-e451-4854-92da-128a2fb99283 req-4722f54f-4821-4ae7-956b-5b87ca0b280c service nova] Releasing lock "refresh_cache-0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.428024] env[62569]: DEBUG nova.compute.manager [req-57b250fa-e451-4854-92da-128a2fb99283 req-4722f54f-4821-4ae7-956b-5b87ca0b280c service nova] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Received event network-changed-6a7245e5-dc62-4799-b00a-34cf50cd5a74 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 777.428024] env[62569]: DEBUG nova.compute.manager [req-57b250fa-e451-4854-92da-128a2fb99283 req-4722f54f-4821-4ae7-956b-5b87ca0b280c service nova] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Refreshing instance network info cache due to event network-changed-6a7245e5-dc62-4799-b00a-34cf50cd5a74. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 777.428024] env[62569]: DEBUG oslo_concurrency.lockutils [req-57b250fa-e451-4854-92da-128a2fb99283 req-4722f54f-4821-4ae7-956b-5b87ca0b280c service nova] Acquiring lock "refresh_cache-77a1b192-6aff-4fee-93d7-57cebcdce626" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.428024] env[62569]: DEBUG oslo_concurrency.lockutils [req-57b250fa-e451-4854-92da-128a2fb99283 req-4722f54f-4821-4ae7-956b-5b87ca0b280c service nova] Acquired lock "refresh_cache-77a1b192-6aff-4fee-93d7-57cebcdce626" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.428385] env[62569]: DEBUG nova.network.neutron [req-57b250fa-e451-4854-92da-128a2fb99283 req-4722f54f-4821-4ae7-956b-5b87ca0b280c service nova] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Refreshing network info cache for port 6a7245e5-dc62-4799-b00a-34cf50cd5a74 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 777.428970] env[62569]: DEBUG oslo_concurrency.lockutils [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Acquired lock "refresh_cache-0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.429281] env[62569]: DEBUG nova.network.neutron [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 777.442357] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.588s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.442357] env[62569]: DEBUG nova.compute.manager [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 777.445981] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.823s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.448195] env[62569]: INFO nova.compute.claims [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 777.565389] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Task: {'id': task-1249977, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.628631] env[62569]: DEBUG nova.network.neutron [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Successfully updated port: cfddbae7-9ab0-4bc6-aea1-46a0cde743fc {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 777.716508] env[62569]: DEBUG oslo_vmware.api [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1249976, 'name': PowerOnVM_Task, 'duration_secs': 0.862996} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.716799] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 777.716986] env[62569]: INFO nova.compute.manager [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Took 8.36 seconds to spawn the instance on the hypervisor. [ 777.717180] env[62569]: DEBUG nova.compute.manager [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 777.717942] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a19284d-1d05-4608-8490-bead3fe74561 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.953271] env[62569]: DEBUG nova.compute.utils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 777.954721] env[62569]: DEBUG nova.compute.manager [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 777.954896] env[62569]: DEBUG nova.network.neutron [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 777.982494] env[62569]: DEBUG nova.network.neutron [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.017706] env[62569]: DEBUG nova.policy [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd72d243a83614002a7e300dc462c7b77', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58f52ff5374a4d5d9969aa2f48f01c8f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 778.068451] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Task: {'id': task-1249977, 'name': ReconfigVM_Task, 'duration_secs': 0.847098} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.069033] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Reconfigured VM instance instance-0000003b to attach disk [datastore1] 4feb7b78-9f7c-4e64-b0a7-870ed73adf97/4feb7b78-9f7c-4e64-b0a7-870ed73adf97.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 778.069771] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a7744adb-31f4-4ca9-9a62-e2d76a56b895 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.079021] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Waiting for the task: (returnval){ [ 778.079021] env[62569]: value = "task-1249978" [ 778.079021] env[62569]: _type = "Task" [ 778.079021] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.085252] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Task: {'id': task-1249978, 'name': Rename_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.131123] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.131507] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquired lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.131849] env[62569]: DEBUG nova.network.neutron [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 778.240610] env[62569]: INFO nova.compute.manager [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Took 35.53 seconds to build instance. [ 778.252298] env[62569]: DEBUG nova.network.neutron [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Updating instance_info_cache with network_info: [{"id": "62cedd28-f07d-4ca0-89b4-ffbc290ff45f", "address": "fa:16:3e:8c:a3:8f", "network": {"id": "327cf78e-624e-4b1a-b3b8-f55c8c47e7c8", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1948247856-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe13e05a444f4f1d94ce9ca68153703e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62cedd28-f0", "ovs_interfaceid": "62cedd28-f07d-4ca0-89b4-ffbc290ff45f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.279630] env[62569]: DEBUG nova.network.neutron [req-57b250fa-e451-4854-92da-128a2fb99283 req-4722f54f-4821-4ae7-956b-5b87ca0b280c service nova] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Updated VIF entry in instance network info cache for port 6a7245e5-dc62-4799-b00a-34cf50cd5a74. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 778.279963] env[62569]: DEBUG nova.network.neutron [req-57b250fa-e451-4854-92da-128a2fb99283 req-4722f54f-4821-4ae7-956b-5b87ca0b280c service nova] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Updating instance_info_cache with network_info: [{"id": "6a7245e5-dc62-4799-b00a-34cf50cd5a74", "address": "fa:16:3e:6f:52:54", "network": {"id": "63365435-28cb-4c92-a35f-dad6842904f6", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-620879956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "548d3ffbc4ac4378a05a8adf0d0a43e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4aa1eda7-48b9-4fa2-af0b-94c718313af2", "external-id": "nsx-vlan-transportzone-502", "segmentation_id": 502, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a7245e5-dc", "ovs_interfaceid": "6a7245e5-dc62-4799-b00a-34cf50cd5a74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.459239] env[62569]: DEBUG nova.compute.manager [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 778.526213] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17b85dd0-b7e9-40ea-91f5-cbea9c56bd0e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.535685] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-48667d58-ecb1-49c4-ac3f-1012cab4fe14 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Suspending the VM {{(pid=62569) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 778.535949] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-4888eeb1-b678-4218-a7db-986c8f697313 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.545173] env[62569]: DEBUG oslo_vmware.api [None req-48667d58-ecb1-49c4-ac3f-1012cab4fe14 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 778.545173] env[62569]: value = "task-1249979" [ 778.545173] env[62569]: _type = "Task" [ 778.545173] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.552759] env[62569]: DEBUG oslo_vmware.api [None req-48667d58-ecb1-49c4-ac3f-1012cab4fe14 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1249979, 'name': SuspendVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.585380] env[62569]: DEBUG nova.compute.manager [req-40d792df-6c67-4f96-a030-c59e87f9e8f4 req-284a817e-a553-4ff1-974b-9d7dc4526c93 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Received event network-vif-plugged-cfddbae7-9ab0-4bc6-aea1-46a0cde743fc {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 778.585816] env[62569]: DEBUG oslo_concurrency.lockutils [req-40d792df-6c67-4f96-a030-c59e87f9e8f4 req-284a817e-a553-4ff1-974b-9d7dc4526c93 service nova] Acquiring lock "b6e5eefc-8c06-445b-a3af-9404578b6179-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.585911] env[62569]: DEBUG oslo_concurrency.lockutils [req-40d792df-6c67-4f96-a030-c59e87f9e8f4 req-284a817e-a553-4ff1-974b-9d7dc4526c93 service nova] Lock "b6e5eefc-8c06-445b-a3af-9404578b6179-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.586085] env[62569]: DEBUG oslo_concurrency.lockutils [req-40d792df-6c67-4f96-a030-c59e87f9e8f4 req-284a817e-a553-4ff1-974b-9d7dc4526c93 service nova] Lock "b6e5eefc-8c06-445b-a3af-9404578b6179-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.588318] env[62569]: DEBUG nova.compute.manager [req-40d792df-6c67-4f96-a030-c59e87f9e8f4 req-284a817e-a553-4ff1-974b-9d7dc4526c93 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] No waiting events found dispatching network-vif-plugged-cfddbae7-9ab0-4bc6-aea1-46a0cde743fc {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 778.588318] env[62569]: WARNING nova.compute.manager [req-40d792df-6c67-4f96-a030-c59e87f9e8f4 req-284a817e-a553-4ff1-974b-9d7dc4526c93 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Received unexpected event network-vif-plugged-cfddbae7-9ab0-4bc6-aea1-46a0cde743fc for instance with vm_state building and task_state spawning. [ 778.588318] env[62569]: DEBUG nova.compute.manager [req-40d792df-6c67-4f96-a030-c59e87f9e8f4 req-284a817e-a553-4ff1-974b-9d7dc4526c93 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Received event network-changed-cfddbae7-9ab0-4bc6-aea1-46a0cde743fc {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 778.588318] env[62569]: DEBUG nova.compute.manager [req-40d792df-6c67-4f96-a030-c59e87f9e8f4 req-284a817e-a553-4ff1-974b-9d7dc4526c93 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Refreshing instance network info cache due to event network-changed-cfddbae7-9ab0-4bc6-aea1-46a0cde743fc. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 778.588318] env[62569]: DEBUG oslo_concurrency.lockutils [req-40d792df-6c67-4f96-a030-c59e87f9e8f4 req-284a817e-a553-4ff1-974b-9d7dc4526c93 service nova] Acquiring lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.592751] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Task: {'id': task-1249978, 'name': Rename_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.661718] env[62569]: DEBUG nova.network.neutron [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Successfully created port: 2a1e0442-6ff2-4954-8798-fd35ba7b59d5 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 778.674326] env[62569]: DEBUG nova.network.neutron [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.744810] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9027fb41-d7f6-4ddb-b0ce-d99250b1ccc0 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "22094c32-5f50-4f86-a77b-cd4adcf8998a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 138.587s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.754467] env[62569]: DEBUG oslo_concurrency.lockutils [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Releasing lock "refresh_cache-0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.754890] env[62569]: DEBUG nova.compute.manager [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Instance network_info: |[{"id": "62cedd28-f07d-4ca0-89b4-ffbc290ff45f", "address": "fa:16:3e:8c:a3:8f", "network": {"id": "327cf78e-624e-4b1a-b3b8-f55c8c47e7c8", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1948247856-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe13e05a444f4f1d94ce9ca68153703e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d903c404-a23a-40c0-a217-96d4bb2e5b08", "external-id": "nsx-vlan-transportzone-228", "segmentation_id": 228, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62cedd28-f0", "ovs_interfaceid": "62cedd28-f07d-4ca0-89b4-ffbc290ff45f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 778.755297] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8c:a3:8f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd903c404-a23a-40c0-a217-96d4bb2e5b08', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '62cedd28-f07d-4ca0-89b4-ffbc290ff45f', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 778.763342] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Creating folder: Project (fe13e05a444f4f1d94ce9ca68153703e). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 778.766802] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e22d9e1e-b1a2-437b-8435-799af7bd940d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.778306] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Created folder: Project (fe13e05a444f4f1d94ce9ca68153703e) in parent group-v269330. [ 778.778522] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Creating folder: Instances. Parent ref: group-v269380. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 778.778800] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a7307d1f-6711-48eb-af8f-a6d6d0b975e7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.783338] env[62569]: DEBUG oslo_concurrency.lockutils [req-57b250fa-e451-4854-92da-128a2fb99283 req-4722f54f-4821-4ae7-956b-5b87ca0b280c service nova] Releasing lock "refresh_cache-77a1b192-6aff-4fee-93d7-57cebcdce626" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.787688] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Created folder: Instances in parent group-v269380. [ 778.787911] env[62569]: DEBUG oslo.service.loopingcall [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 778.788113] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 778.788348] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6146b203-63a0-4a11-82ac-69a3736d6e9c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.809715] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 778.809715] env[62569]: value = "task-1249982" [ 778.809715] env[62569]: _type = "Task" [ 778.809715] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.816992] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249982, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.819807] env[62569]: DEBUG nova.network.neutron [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Updating instance_info_cache with network_info: [{"id": "cfddbae7-9ab0-4bc6-aea1-46a0cde743fc", "address": "fa:16:3e:b7:d6:80", "network": {"id": "406dd658-8168-4e73-9b97-861929bfaa2e", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-693938985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de8605118d744a93bce54e897aa849b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfddbae7-9a", "ovs_interfaceid": "cfddbae7-9ab0-4bc6-aea1-46a0cde743fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.843499] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2540435-662b-4d73-ab7b-8c4dcb48cd78 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.851797] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-658d90e9-c441-4408-8d56-1e7d3ffe97b1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.885448] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1bda4c8-d3a0-498b-919a-7920a39dd580 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.893638] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35bf85d9-43ad-4daa-a22c-1b3612c5e11f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.908503] env[62569]: DEBUG nova.compute.provider_tree [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 779.054521] env[62569]: DEBUG oslo_vmware.api [None req-48667d58-ecb1-49c4-ac3f-1012cab4fe14 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1249979, 'name': SuspendVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.087422] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Task: {'id': task-1249978, 'name': Rename_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.246274] env[62569]: DEBUG nova.compute.manager [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 779.319980] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249982, 'name': CreateVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.321978] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Releasing lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.322335] env[62569]: DEBUG nova.compute.manager [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Instance network_info: |[{"id": "cfddbae7-9ab0-4bc6-aea1-46a0cde743fc", "address": "fa:16:3e:b7:d6:80", "network": {"id": "406dd658-8168-4e73-9b97-861929bfaa2e", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-693938985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de8605118d744a93bce54e897aa849b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfddbae7-9a", "ovs_interfaceid": "cfddbae7-9ab0-4bc6-aea1-46a0cde743fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 779.322682] env[62569]: DEBUG oslo_concurrency.lockutils [req-40d792df-6c67-4f96-a030-c59e87f9e8f4 req-284a817e-a553-4ff1-974b-9d7dc4526c93 service nova] Acquired lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.322886] env[62569]: DEBUG nova.network.neutron [req-40d792df-6c67-4f96-a030-c59e87f9e8f4 req-284a817e-a553-4ff1-974b-9d7dc4526c93 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Refreshing network info cache for port cfddbae7-9ab0-4bc6-aea1-46a0cde743fc {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 779.324237] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b7:d6:80', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '257e5ea7-8b80-4301-9900-a754f1fe2031', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cfddbae7-9ab0-4bc6-aea1-46a0cde743fc', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 779.333242] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Creating folder: Project (de8605118d744a93bce54e897aa849b5). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 779.333844] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d11ce0ca-9043-4536-a30e-ac5386106a36 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.344371] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Created folder: Project (de8605118d744a93bce54e897aa849b5) in parent group-v269330. [ 779.344543] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Creating folder: Instances. Parent ref: group-v269383. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 779.345011] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5f80fc01-9162-4f90-855e-cf5966bb3b24 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.354630] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Created folder: Instances in parent group-v269383. [ 779.354876] env[62569]: DEBUG oslo.service.loopingcall [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 779.355082] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 779.355296] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e5ca216-066d-462d-9e76-e481491f57ad {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.374412] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 779.374412] env[62569]: value = "task-1249985" [ 779.374412] env[62569]: _type = "Task" [ 779.374412] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.382680] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249985, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.412272] env[62569]: DEBUG nova.scheduler.client.report [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 779.468685] env[62569]: DEBUG nova.compute.manager [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 779.494797] env[62569]: DEBUG nova.virt.hardware [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 779.494797] env[62569]: DEBUG nova.virt.hardware [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 779.494797] env[62569]: DEBUG nova.virt.hardware [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 779.495023] env[62569]: DEBUG nova.virt.hardware [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 779.495023] env[62569]: DEBUG nova.virt.hardware [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 779.495023] env[62569]: DEBUG nova.virt.hardware [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 779.495180] env[62569]: DEBUG nova.virt.hardware [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 779.495338] env[62569]: DEBUG nova.virt.hardware [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 779.495504] env[62569]: DEBUG nova.virt.hardware [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 779.495664] env[62569]: DEBUG nova.virt.hardware [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 779.495858] env[62569]: DEBUG nova.virt.hardware [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 779.497105] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2a97e41-3aef-43bd-9326-2b158fedb812 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.505534] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b38b4b-f07d-4437-9136-c089d523cbc0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.555569] env[62569]: DEBUG oslo_vmware.api [None req-48667d58-ecb1-49c4-ac3f-1012cab4fe14 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1249979, 'name': SuspendVM_Task} progress is 58%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.588249] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Task: {'id': task-1249978, 'name': Rename_Task, 'duration_secs': 1.170144} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.588537] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 779.588788] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4870fc49-0820-40a0-bf96-bd58742ee85e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.596028] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Waiting for the task: (returnval){ [ 779.596028] env[62569]: value = "task-1249986" [ 779.596028] env[62569]: _type = "Task" [ 779.596028] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.604540] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Task: {'id': task-1249986, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.771707] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.822847] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249982, 'name': CreateVM_Task, 'duration_secs': 0.536195} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.823110] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 779.823969] env[62569]: DEBUG oslo_concurrency.lockutils [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.824209] env[62569]: DEBUG oslo_concurrency.lockutils [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.824643] env[62569]: DEBUG oslo_concurrency.lockutils [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 779.824964] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82adf3be-85be-455c-8756-eeee8eb970bb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.830696] env[62569]: DEBUG oslo_vmware.api [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Waiting for the task: (returnval){ [ 779.830696] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52052654-cf24-3c52-b262-5dbe9912cfb7" [ 779.830696] env[62569]: _type = "Task" [ 779.830696] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.845840] env[62569]: DEBUG oslo_vmware.api [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52052654-cf24-3c52-b262-5dbe9912cfb7, 'name': SearchDatastore_Task, 'duration_secs': 0.009775} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.846205] env[62569]: DEBUG oslo_concurrency.lockutils [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.846545] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 779.846881] env[62569]: DEBUG oslo_concurrency.lockutils [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.847105] env[62569]: DEBUG oslo_concurrency.lockutils [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.847347] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 779.847686] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-060659cb-b798-4bf2-9fff-d2b2b3b08748 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.856490] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 779.856726] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 779.857533] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7240e5e6-9d47-4802-9df3-eb494dde3204 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.863266] env[62569]: DEBUG oslo_vmware.api [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Waiting for the task: (returnval){ [ 779.863266] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ca54f5-eea1-2d46-3e98-b0cb6ca3d778" [ 779.863266] env[62569]: _type = "Task" [ 779.863266] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.874820] env[62569]: DEBUG oslo_vmware.api [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ca54f5-eea1-2d46-3e98-b0cb6ca3d778, 'name': SearchDatastore_Task, 'duration_secs': 0.008287} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.878891] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b28d46eb-5e5e-4aa7-80ea-3b6afa43a875 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.887131] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249985, 'name': CreateVM_Task, 'duration_secs': 0.450898} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.887925] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 779.888273] env[62569]: DEBUG oslo_vmware.api [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Waiting for the task: (returnval){ [ 779.888273] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5236da95-d6e7-757b-980f-199ee7891fef" [ 779.888273] env[62569]: _type = "Task" [ 779.888273] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.888922] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.889102] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.889427] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 779.889728] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b050a1c1-26b0-4b25-882b-106f2f0ddeaa {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.899912] env[62569]: DEBUG oslo_vmware.api [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5236da95-d6e7-757b-980f-199ee7891fef, 'name': SearchDatastore_Task, 'duration_secs': 0.009151} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.900770] env[62569]: DEBUG oslo_concurrency.lockutils [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.901085] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010/0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 779.901382] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 779.901382] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5241c6a2-cd63-af13-a3ab-1003bdabf3c9" [ 779.901382] env[62569]: _type = "Task" [ 779.901382] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.901556] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-78b2451b-24d4-4a3a-8e02-fcc2717cf51b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.913962] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5241c6a2-cd63-af13-a3ab-1003bdabf3c9, 'name': SearchDatastore_Task, 'duration_secs': 0.009276} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.915226] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.915465] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 779.915693] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.915861] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.916083] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 779.916390] env[62569]: DEBUG oslo_vmware.api [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Waiting for the task: (returnval){ [ 779.916390] env[62569]: value = "task-1249987" [ 779.916390] env[62569]: _type = "Task" [ 779.916390] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.917235] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.471s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.917580] env[62569]: DEBUG nova.compute.manager [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 779.920151] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-356881d0-ec1d-4b79-a46c-aef22de9b0db {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.924610] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.364s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.924610] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.928282] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.159s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.932191] env[62569]: INFO nova.compute.claims [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 779.944631] env[62569]: DEBUG oslo_vmware.api [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Task: {'id': task-1249987, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.955350] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 779.955524] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 779.956271] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6edf7f6d-6c19-4ba9-b97c-1a9f1d13396f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.964959] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 779.964959] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5249bc94-4724-f819-bcc4-40356bdf0a50" [ 779.964959] env[62569]: _type = "Task" [ 779.964959] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.966033] env[62569]: INFO nova.scheduler.client.report [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Deleted allocations for instance 3f504d5c-3431-4a80-a7d9-e52af01cf24e [ 779.982025] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5249bc94-4724-f819-bcc4-40356bdf0a50, 'name': SearchDatastore_Task, 'duration_secs': 0.008821} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.982025] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-172b7774-b565-4f54-b0de-2e92d4e1342f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.988258] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 779.988258] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52fed04c-2025-3dc7-e112-e25ce94f9f37" [ 779.988258] env[62569]: _type = "Task" [ 779.988258] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.001247] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52fed04c-2025-3dc7-e112-e25ce94f9f37, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.056616] env[62569]: DEBUG oslo_vmware.api [None req-48667d58-ecb1-49c4-ac3f-1012cab4fe14 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1249979, 'name': SuspendVM_Task, 'duration_secs': 1.336091} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.057075] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-48667d58-ecb1-49c4-ac3f-1012cab4fe14 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Suspended the VM {{(pid=62569) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 780.057395] env[62569]: DEBUG nova.compute.manager [None req-48667d58-ecb1-49c4-ac3f-1012cab4fe14 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 780.059647] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03aade6d-2721-4007-b8f7-13e7c6670033 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.109959] env[62569]: DEBUG oslo_vmware.api [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Task: {'id': task-1249986, 'name': PowerOnVM_Task, 'duration_secs': 0.494711} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.110483] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 780.110827] env[62569]: INFO nova.compute.manager [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Took 8.24 seconds to spawn the instance on the hypervisor. [ 780.111163] env[62569]: DEBUG nova.compute.manager [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 780.112416] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-294ba492-7584-49be-b6b8-7697ef28ab81 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.281719] env[62569]: DEBUG nova.network.neutron [req-40d792df-6c67-4f96-a030-c59e87f9e8f4 req-284a817e-a553-4ff1-974b-9d7dc4526c93 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Updated VIF entry in instance network info cache for port cfddbae7-9ab0-4bc6-aea1-46a0cde743fc. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 780.282121] env[62569]: DEBUG nova.network.neutron [req-40d792df-6c67-4f96-a030-c59e87f9e8f4 req-284a817e-a553-4ff1-974b-9d7dc4526c93 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Updating instance_info_cache with network_info: [{"id": "cfddbae7-9ab0-4bc6-aea1-46a0cde743fc", "address": "fa:16:3e:b7:d6:80", "network": {"id": "406dd658-8168-4e73-9b97-861929bfaa2e", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-693938985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de8605118d744a93bce54e897aa849b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfddbae7-9a", "ovs_interfaceid": "cfddbae7-9ab0-4bc6-aea1-46a0cde743fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.423767] env[62569]: DEBUG nova.compute.utils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 780.434836] env[62569]: DEBUG nova.compute.manager [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 780.435032] env[62569]: DEBUG nova.network.neutron [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 780.445734] env[62569]: DEBUG oslo_vmware.api [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Task: {'id': task-1249987, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.466926} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.446126] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010/0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 780.446528] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 780.447068] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5bb4a060-32c9-4b58-a4c5-856e3e35a123 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.454597] env[62569]: DEBUG oslo_vmware.api [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Waiting for the task: (returnval){ [ 780.454597] env[62569]: value = "task-1249988" [ 780.454597] env[62569]: _type = "Task" [ 780.454597] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.463616] env[62569]: DEBUG oslo_vmware.api [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Task: {'id': task-1249988, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.470508] env[62569]: DEBUG nova.compute.manager [req-2d20343e-0a87-4a21-b708-5181143c8406 req-be713fb4-eba7-40d1-a7e8-832c0b728aeb service nova] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Received event network-vif-plugged-2a1e0442-6ff2-4954-8798-fd35ba7b59d5 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 780.470734] env[62569]: DEBUG oslo_concurrency.lockutils [req-2d20343e-0a87-4a21-b708-5181143c8406 req-be713fb4-eba7-40d1-a7e8-832c0b728aeb service nova] Acquiring lock "37a4c60a-e660-4c91-bcfe-72638a4397b0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.470990] env[62569]: DEBUG oslo_concurrency.lockutils [req-2d20343e-0a87-4a21-b708-5181143c8406 req-be713fb4-eba7-40d1-a7e8-832c0b728aeb service nova] Lock "37a4c60a-e660-4c91-bcfe-72638a4397b0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.471205] env[62569]: DEBUG oslo_concurrency.lockutils [req-2d20343e-0a87-4a21-b708-5181143c8406 req-be713fb4-eba7-40d1-a7e8-832c0b728aeb service nova] Lock "37a4c60a-e660-4c91-bcfe-72638a4397b0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.471363] env[62569]: DEBUG nova.compute.manager [req-2d20343e-0a87-4a21-b708-5181143c8406 req-be713fb4-eba7-40d1-a7e8-832c0b728aeb service nova] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] No waiting events found dispatching network-vif-plugged-2a1e0442-6ff2-4954-8798-fd35ba7b59d5 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 780.471570] env[62569]: WARNING nova.compute.manager [req-2d20343e-0a87-4a21-b708-5181143c8406 req-be713fb4-eba7-40d1-a7e8-832c0b728aeb service nova] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Received unexpected event network-vif-plugged-2a1e0442-6ff2-4954-8798-fd35ba7b59d5 for instance with vm_state building and task_state spawning. [ 780.477824] env[62569]: DEBUG nova.policy [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd72d243a83614002a7e300dc462c7b77', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58f52ff5374a4d5d9969aa2f48f01c8f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 780.479664] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ba7342fa-2a52-4752-b3d3-b6062a730bdd tempest-TenantUsagesTestJSON-670341100 tempest-TenantUsagesTestJSON-670341100-project-member] Lock "3f504d5c-3431-4a80-a7d9-e52af01cf24e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.495s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.498624] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52fed04c-2025-3dc7-e112-e25ce94f9f37, 'name': SearchDatastore_Task, 'duration_secs': 0.017778} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.499455] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.499715] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] b6e5eefc-8c06-445b-a3af-9404578b6179/b6e5eefc-8c06-445b-a3af-9404578b6179.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 780.499967] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3b930045-fa3c-40a8-9a7b-bba234cadf6a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.507959] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 780.507959] env[62569]: value = "task-1249989" [ 780.507959] env[62569]: _type = "Task" [ 780.507959] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.516218] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1249989, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.599046] env[62569]: DEBUG nova.network.neutron [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Successfully updated port: 2a1e0442-6ff2-4954-8798-fd35ba7b59d5 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 780.639802] env[62569]: INFO nova.compute.manager [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Took 32.17 seconds to build instance. [ 780.765079] env[62569]: DEBUG nova.network.neutron [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Successfully created port: 1ae266f0-188a-45de-9a09-d9e21b48bff3 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 780.785726] env[62569]: DEBUG oslo_concurrency.lockutils [req-40d792df-6c67-4f96-a030-c59e87f9e8f4 req-284a817e-a553-4ff1-974b-9d7dc4526c93 service nova] Releasing lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.938581] env[62569]: DEBUG nova.compute.manager [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 780.964719] env[62569]: DEBUG oslo_vmware.api [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Task: {'id': task-1249988, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061198} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.965020] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 780.965866] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e7fbfbe-567d-42da-b56d-c5b7dcc5540d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.994335] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010/0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 780.998301] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a6024a3-1295-4857-8375-a3d0021284d6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.023808] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1249989, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.434718} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.025208] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] b6e5eefc-8c06-445b-a3af-9404578b6179/b6e5eefc-8c06-445b-a3af-9404578b6179.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 781.025357] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 781.025666] env[62569]: DEBUG oslo_vmware.api [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Waiting for the task: (returnval){ [ 781.025666] env[62569]: value = "task-1249990" [ 781.025666] env[62569]: _type = "Task" [ 781.025666] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.025856] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d13d9014-4cf8-46f5-b1dc-a27f4e0a1369 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.040018] env[62569]: DEBUG oslo_vmware.api [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Task: {'id': task-1249990, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.040269] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 781.040269] env[62569]: value = "task-1249991" [ 781.040269] env[62569]: _type = "Task" [ 781.040269] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.051020] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1249991, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.101692] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "refresh_cache-37a4c60a-e660-4c91-bcfe-72638a4397b0" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.101692] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquired lock "refresh_cache-37a4c60a-e660-4c91-bcfe-72638a4397b0" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.102122] env[62569]: DEBUG nova.network.neutron [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 781.142912] env[62569]: DEBUG oslo_concurrency.lockutils [None req-764e32f4-5322-412d-80a0-d96f6d1287d5 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Lock "4feb7b78-9f7c-4e64-b0a7-870ed73adf97" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.359s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.352167] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d223690-67e8-485e-8e3a-755856d04a6c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.362853] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c19fc329-ea94-443b-b10e-a48495bf2def {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.394704] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-064751be-8183-4dc5-8818-57afb3a4cafd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.402097] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cf23051-582a-401d-a8c5-262ba48a7559 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.416344] env[62569]: DEBUG nova.compute.provider_tree [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 781.537126] env[62569]: DEBUG oslo_vmware.api [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Task: {'id': task-1249990, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.547682] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1249991, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.234894} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.548048] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 781.549242] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6aa5029-a099-4d1a-9ffb-8385d0220454 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.573958] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] b6e5eefc-8c06-445b-a3af-9404578b6179/b6e5eefc-8c06-445b-a3af-9404578b6179.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 781.574458] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59376f26-eb5c-46b6-931e-2e43029f6a78 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.596381] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 781.596381] env[62569]: value = "task-1249992" [ 781.596381] env[62569]: _type = "Task" [ 781.596381] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.609680] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1249992, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.645341] env[62569]: DEBUG nova.compute.manager [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 781.652066] env[62569]: DEBUG nova.network.neutron [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 781.858274] env[62569]: DEBUG nova.network.neutron [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Updating instance_info_cache with network_info: [{"id": "2a1e0442-6ff2-4954-8798-fd35ba7b59d5", "address": "fa:16:3e:0e:d4:4d", "network": {"id": "f2b76a4d-be64-4c9e-8c25-d2bbe42f59ea", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1621748668-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58f52ff5374a4d5d9969aa2f48f01c8f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2907cce-d529-4809-af05-d29397bed211", "external-id": "nsx-vlan-transportzone-427", "segmentation_id": 427, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a1e0442-6f", "ovs_interfaceid": "2a1e0442-6ff2-4954-8798-fd35ba7b59d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.919824] env[62569]: DEBUG nova.scheduler.client.report [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 781.950425] env[62569]: DEBUG nova.compute.manager [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 781.980351] env[62569]: DEBUG nova.virt.hardware [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 781.981082] env[62569]: DEBUG nova.virt.hardware [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 781.981082] env[62569]: DEBUG nova.virt.hardware [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 781.981444] env[62569]: DEBUG nova.virt.hardware [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 781.982077] env[62569]: DEBUG nova.virt.hardware [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 781.982284] env[62569]: DEBUG nova.virt.hardware [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 781.982514] env[62569]: DEBUG nova.virt.hardware [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 781.982723] env[62569]: DEBUG nova.virt.hardware [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 781.984046] env[62569]: DEBUG nova.virt.hardware [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 781.984046] env[62569]: DEBUG nova.virt.hardware [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 781.984046] env[62569]: DEBUG nova.virt.hardware [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 781.984918] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fea1ef8e-c6c9-4504-8054-878ab0b00ba6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.993023] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cc17fd0-9342-4309-a541-694c095e0744 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.038174] env[62569]: DEBUG oslo_vmware.api [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Task: {'id': task-1249990, 'name': ReconfigVM_Task, 'duration_secs': 0.720357} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.038174] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010/0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 782.040702] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-90364ced-970b-4e67-8539-be8beb26bd9d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.047042] env[62569]: DEBUG oslo_vmware.api [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Waiting for the task: (returnval){ [ 782.047042] env[62569]: value = "task-1249993" [ 782.047042] env[62569]: _type = "Task" [ 782.047042] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.055150] env[62569]: DEBUG oslo_vmware.api [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Task: {'id': task-1249993, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.107955] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1249992, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.179653] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.270553] env[62569]: DEBUG nova.compute.manager [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 782.271570] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44b82af2-09d6-4977-8016-c3505dfb81be {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.361841] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Releasing lock "refresh_cache-37a4c60a-e660-4c91-bcfe-72638a4397b0" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.362260] env[62569]: DEBUG nova.compute.manager [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Instance network_info: |[{"id": "2a1e0442-6ff2-4954-8798-fd35ba7b59d5", "address": "fa:16:3e:0e:d4:4d", "network": {"id": "f2b76a4d-be64-4c9e-8c25-d2bbe42f59ea", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1621748668-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58f52ff5374a4d5d9969aa2f48f01c8f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2907cce-d529-4809-af05-d29397bed211", "external-id": "nsx-vlan-transportzone-427", "segmentation_id": 427, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a1e0442-6f", "ovs_interfaceid": "2a1e0442-6ff2-4954-8798-fd35ba7b59d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 782.363283] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:d4:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b2907cce-d529-4809-af05-d29397bed211', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2a1e0442-6ff2-4954-8798-fd35ba7b59d5', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 782.371900] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Creating folder: Project (58f52ff5374a4d5d9969aa2f48f01c8f). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 782.372209] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-da12d625-7da2-4f21-8e29-1664b315ca67 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.385047] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Created folder: Project (58f52ff5374a4d5d9969aa2f48f01c8f) in parent group-v269330. [ 782.385255] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Creating folder: Instances. Parent ref: group-v269386. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 782.385482] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c85780ac-b7ac-47a7-a156-e2f8c60b5132 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.393834] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Created folder: Instances in parent group-v269386. [ 782.394033] env[62569]: DEBUG oslo.service.loopingcall [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 782.394266] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 782.394426] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8b7a8cd2-3fbf-4b55-adf1-e2b5d6fb98fd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.415122] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 782.415122] env[62569]: value = "task-1249996" [ 782.415122] env[62569]: _type = "Task" [ 782.415122] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.425487] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249996, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.429871] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.502s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.429871] env[62569]: DEBUG nova.compute.manager [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 782.435719] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.443s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.440183] env[62569]: INFO nova.compute.claims [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 782.443659] env[62569]: DEBUG nova.compute.manager [req-bc1f0284-44c1-4a9e-87b2-f2fc9ad7149d req-cce0e3ba-0f77-46cc-9149-1a704033d86d service nova] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Received event network-vif-plugged-1ae266f0-188a-45de-9a09-d9e21b48bff3 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 782.443878] env[62569]: DEBUG oslo_concurrency.lockutils [req-bc1f0284-44c1-4a9e-87b2-f2fc9ad7149d req-cce0e3ba-0f77-46cc-9149-1a704033d86d service nova] Acquiring lock "06067572-a9fd-43a4-91cd-383a62f53885-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.444093] env[62569]: DEBUG oslo_concurrency.lockutils [req-bc1f0284-44c1-4a9e-87b2-f2fc9ad7149d req-cce0e3ba-0f77-46cc-9149-1a704033d86d service nova] Lock "06067572-a9fd-43a4-91cd-383a62f53885-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.444257] env[62569]: DEBUG oslo_concurrency.lockutils [req-bc1f0284-44c1-4a9e-87b2-f2fc9ad7149d req-cce0e3ba-0f77-46cc-9149-1a704033d86d service nova] Lock "06067572-a9fd-43a4-91cd-383a62f53885-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.444420] env[62569]: DEBUG nova.compute.manager [req-bc1f0284-44c1-4a9e-87b2-f2fc9ad7149d req-cce0e3ba-0f77-46cc-9149-1a704033d86d service nova] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] No waiting events found dispatching network-vif-plugged-1ae266f0-188a-45de-9a09-d9e21b48bff3 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 782.444576] env[62569]: WARNING nova.compute.manager [req-bc1f0284-44c1-4a9e-87b2-f2fc9ad7149d req-cce0e3ba-0f77-46cc-9149-1a704033d86d service nova] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Received unexpected event network-vif-plugged-1ae266f0-188a-45de-9a09-d9e21b48bff3 for instance with vm_state building and task_state spawning. [ 782.507244] env[62569]: DEBUG nova.compute.manager [req-4921ba4c-b25e-41af-89ab-694c07094534 req-5c3a3a7d-2672-4170-bd68-1c31c2612283 service nova] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Received event network-changed-2a1e0442-6ff2-4954-8798-fd35ba7b59d5 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 782.507468] env[62569]: DEBUG nova.compute.manager [req-4921ba4c-b25e-41af-89ab-694c07094534 req-5c3a3a7d-2672-4170-bd68-1c31c2612283 service nova] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Refreshing instance network info cache due to event network-changed-2a1e0442-6ff2-4954-8798-fd35ba7b59d5. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 782.507677] env[62569]: DEBUG oslo_concurrency.lockutils [req-4921ba4c-b25e-41af-89ab-694c07094534 req-5c3a3a7d-2672-4170-bd68-1c31c2612283 service nova] Acquiring lock "refresh_cache-37a4c60a-e660-4c91-bcfe-72638a4397b0" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.507828] env[62569]: DEBUG oslo_concurrency.lockutils [req-4921ba4c-b25e-41af-89ab-694c07094534 req-5c3a3a7d-2672-4170-bd68-1c31c2612283 service nova] Acquired lock "refresh_cache-37a4c60a-e660-4c91-bcfe-72638a4397b0" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.507996] env[62569]: DEBUG nova.network.neutron [req-4921ba4c-b25e-41af-89ab-694c07094534 req-5c3a3a7d-2672-4170-bd68-1c31c2612283 service nova] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Refreshing network info cache for port 2a1e0442-6ff2-4954-8798-fd35ba7b59d5 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 782.554504] env[62569]: DEBUG nova.network.neutron [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Successfully updated port: 1ae266f0-188a-45de-9a09-d9e21b48bff3 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 782.558756] env[62569]: DEBUG oslo_vmware.api [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Task: {'id': task-1249993, 'name': Rename_Task, 'duration_secs': 0.130951} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.559263] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 782.559593] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8161cfd4-497d-44ac-9044-0a9617e39d46 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.566902] env[62569]: DEBUG oslo_vmware.api [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Waiting for the task: (returnval){ [ 782.566902] env[62569]: value = "task-1249997" [ 782.566902] env[62569]: _type = "Task" [ 782.566902] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.577666] env[62569]: DEBUG oslo_vmware.api [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Task: {'id': task-1249997, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.608217] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1249992, 'name': ReconfigVM_Task, 'duration_secs': 0.637923} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.608627] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Reconfigured VM instance instance-0000003d to attach disk [datastore2] b6e5eefc-8c06-445b-a3af-9404578b6179/b6e5eefc-8c06-445b-a3af-9404578b6179.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 782.609357] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f5389142-0170-4728-b7ca-a9d4e1d3a081 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.616166] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 782.616166] env[62569]: value = "task-1249998" [ 782.616166] env[62569]: _type = "Task" [ 782.616166] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.625144] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1249998, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.787037] env[62569]: INFO nova.compute.manager [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] instance snapshotting [ 782.787037] env[62569]: WARNING nova.compute.manager [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 782.793218] env[62569]: DEBUG oslo_concurrency.lockutils [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Acquiring lock "4feb7b78-9f7c-4e64-b0a7-870ed73adf97" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.793218] env[62569]: DEBUG oslo_concurrency.lockutils [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Lock "4feb7b78-9f7c-4e64-b0a7-870ed73adf97" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.793218] env[62569]: DEBUG oslo_concurrency.lockutils [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Acquiring lock "4feb7b78-9f7c-4e64-b0a7-870ed73adf97-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.793218] env[62569]: DEBUG oslo_concurrency.lockutils [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Lock "4feb7b78-9f7c-4e64-b0a7-870ed73adf97-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.793560] env[62569]: DEBUG oslo_concurrency.lockutils [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Lock "4feb7b78-9f7c-4e64-b0a7-870ed73adf97-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.795462] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ab5072c-641d-481e-91be-4fdfb6cddfef {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.799844] env[62569]: INFO nova.compute.manager [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Terminating instance [ 782.826208] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a68005c3-9cc7-40b8-8397-b77fec25ee4e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.931703] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1249996, 'name': CreateVM_Task, 'duration_secs': 0.351349} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.931985] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 782.932711] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.932950] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.933333] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 782.933606] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7069b40e-c2f3-445c-8289-8373e2c4bf2b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.936497] env[62569]: DEBUG nova.compute.utils [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 782.938032] env[62569]: DEBUG nova.compute.manager [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 782.938198] env[62569]: DEBUG nova.network.neutron [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 782.943230] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for the task: (returnval){ [ 782.943230] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528c3c1e-8648-4e77-b26c-931ee7b81a3c" [ 782.943230] env[62569]: _type = "Task" [ 782.943230] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.959504] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528c3c1e-8648-4e77-b26c-931ee7b81a3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.005510] env[62569]: DEBUG nova.policy [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '542d863e3c52455781d9bf381c007c7a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7633be3f1b38470dae7048754b408f0c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 783.060714] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "refresh_cache-06067572-a9fd-43a4-91cd-383a62f53885" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.060913] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquired lock "refresh_cache-06067572-a9fd-43a4-91cd-383a62f53885" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.061180] env[62569]: DEBUG nova.network.neutron [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 783.081793] env[62569]: DEBUG oslo_vmware.api [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Task: {'id': task-1249997, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.128115] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1249998, 'name': Rename_Task, 'duration_secs': 0.161275} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.129039] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 783.129458] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cea5cea0-195e-4a60-817a-dbd9ca10e6d6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.139466] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 783.139466] env[62569]: value = "task-1249999" [ 783.139466] env[62569]: _type = "Task" [ 783.139466] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.149449] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1249999, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.295531] env[62569]: DEBUG nova.network.neutron [req-4921ba4c-b25e-41af-89ab-694c07094534 req-5c3a3a7d-2672-4170-bd68-1c31c2612283 service nova] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Updated VIF entry in instance network info cache for port 2a1e0442-6ff2-4954-8798-fd35ba7b59d5. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 783.295531] env[62569]: DEBUG nova.network.neutron [req-4921ba4c-b25e-41af-89ab-694c07094534 req-5c3a3a7d-2672-4170-bd68-1c31c2612283 service nova] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Updating instance_info_cache with network_info: [{"id": "2a1e0442-6ff2-4954-8798-fd35ba7b59d5", "address": "fa:16:3e:0e:d4:4d", "network": {"id": "f2b76a4d-be64-4c9e-8c25-d2bbe42f59ea", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1621748668-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58f52ff5374a4d5d9969aa2f48f01c8f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2907cce-d529-4809-af05-d29397bed211", "external-id": "nsx-vlan-transportzone-427", "segmentation_id": 427, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a1e0442-6f", "ovs_interfaceid": "2a1e0442-6ff2-4954-8798-fd35ba7b59d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.331065] env[62569]: DEBUG nova.compute.manager [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 783.333494] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 783.333494] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd2e1214-9453-4220-93a5-b13771f4e4b8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.339045] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Creating Snapshot of the VM instance {{(pid=62569) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 783.339045] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-11f519e6-85ec-4f46-b000-ef67a2ae9509 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.342377] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 783.343159] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ad0abc0a-d187-4f35-94d2-e78e458541fa {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.347865] env[62569]: DEBUG oslo_vmware.api [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 783.347865] env[62569]: value = "task-1250000" [ 783.347865] env[62569]: _type = "Task" [ 783.347865] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.352012] env[62569]: DEBUG oslo_vmware.api [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Waiting for the task: (returnval){ [ 783.352012] env[62569]: value = "task-1250001" [ 783.352012] env[62569]: _type = "Task" [ 783.352012] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.357845] env[62569]: DEBUG oslo_vmware.api [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250000, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.363506] env[62569]: DEBUG oslo_vmware.api [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Task: {'id': task-1250001, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.441174] env[62569]: DEBUG nova.compute.manager [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 783.456643] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528c3c1e-8648-4e77-b26c-931ee7b81a3c, 'name': SearchDatastore_Task, 'duration_secs': 0.014276} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.457870] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.457870] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 783.457870] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.457870] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.460630] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 783.460630] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d5c47b67-ddc0-49f5-ab1e-0c800161167d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.471495] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 783.471684] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 783.472457] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97d9f135-c149-4245-b371-176085ad36da {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.482580] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for the task: (returnval){ [ 783.482580] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d90409-caf8-5feb-ebae-3d82cd06dbae" [ 783.482580] env[62569]: _type = "Task" [ 783.482580] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.486626] env[62569]: DEBUG nova.network.neutron [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Successfully created port: 06154dac-fb91-4e11-a503-34ae0c293331 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 783.501522] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d90409-caf8-5feb-ebae-3d82cd06dbae, 'name': SearchDatastore_Task, 'duration_secs': 0.011506} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.502304] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7198528-1ea4-41b9-8854-4e2edda37322 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.508508] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for the task: (returnval){ [ 783.508508] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]521352bc-bed2-beb3-4755-66424f297c75" [ 783.508508] env[62569]: _type = "Task" [ 783.508508] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.523818] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]521352bc-bed2-beb3-4755-66424f297c75, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.583032] env[62569]: DEBUG oslo_vmware.api [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Task: {'id': task-1249997, 'name': PowerOnVM_Task, 'duration_secs': 0.516816} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.583407] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 783.583775] env[62569]: INFO nova.compute.manager [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Took 9.25 seconds to spawn the instance on the hypervisor. [ 783.583775] env[62569]: DEBUG nova.compute.manager [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 783.590537] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-389f7b47-8f94-4d6e-ad41-e20548597ed0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.593958] env[62569]: DEBUG nova.network.neutron [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.654383] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1249999, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.798172] env[62569]: DEBUG oslo_concurrency.lockutils [req-4921ba4c-b25e-41af-89ab-694c07094534 req-5c3a3a7d-2672-4170-bd68-1c31c2612283 service nova] Releasing lock "refresh_cache-37a4c60a-e660-4c91-bcfe-72638a4397b0" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.862608] env[62569]: DEBUG oslo_vmware.api [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250000, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.867510] env[62569]: DEBUG oslo_vmware.api [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Task: {'id': task-1250001, 'name': PowerOffVM_Task, 'duration_secs': 0.21828} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.867894] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 783.868169] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 783.870644] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f307bb87-a411-4ef5-b86b-8896f3fbdcb1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.897569] env[62569]: DEBUG nova.network.neutron [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Updating instance_info_cache with network_info: [{"id": "1ae266f0-188a-45de-9a09-d9e21b48bff3", "address": "fa:16:3e:6b:d1:2a", "network": {"id": "f2b76a4d-be64-4c9e-8c25-d2bbe42f59ea", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1621748668-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58f52ff5374a4d5d9969aa2f48f01c8f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2907cce-d529-4809-af05-d29397bed211", "external-id": "nsx-vlan-transportzone-427", "segmentation_id": 427, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1ae266f0-18", "ovs_interfaceid": "1ae266f0-188a-45de-9a09-d9e21b48bff3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.940716] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 783.941064] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 783.941436] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Deleting the datastore file [datastore1] 4feb7b78-9f7c-4e64-b0a7-870ed73adf97 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 783.941824] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-54006c7e-0e66-4073-9468-4bb467a4ecbc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.954154] env[62569]: DEBUG oslo_vmware.api [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Waiting for the task: (returnval){ [ 783.954154] env[62569]: value = "task-1250003" [ 783.954154] env[62569]: _type = "Task" [ 783.954154] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.964970] env[62569]: DEBUG oslo_vmware.api [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Task: {'id': task-1250003, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.990105] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6f26373-306e-4f3e-a942-1df16a72687c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.999824] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0872ad05-6fdf-4448-9a32-836ed95bf45f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.039019] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b85915ec-33a8-4e33-baab-0bd74bd2a731 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.048374] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]521352bc-bed2-beb3-4755-66424f297c75, 'name': SearchDatastore_Task, 'duration_secs': 0.012163} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.050568] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.050888] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 37a4c60a-e660-4c91-bcfe-72638a4397b0/37a4c60a-e660-4c91-bcfe-72638a4397b0.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 784.051157] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3cd11632-da29-483d-9cd4-76bbb323d7d8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.053955] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e94b77cc-ac07-4905-b74a-37572512a347 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.069242] env[62569]: DEBUG nova.compute.provider_tree [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.071024] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for the task: (returnval){ [ 784.071024] env[62569]: value = "task-1250004" [ 784.071024] env[62569]: _type = "Task" [ 784.071024] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.079928] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250004, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.111091] env[62569]: INFO nova.compute.manager [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Took 34.05 seconds to build instance. [ 784.151474] env[62569]: DEBUG oslo_vmware.api [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1249999, 'name': PowerOnVM_Task, 'duration_secs': 0.787449} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.151786] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 784.151986] env[62569]: INFO nova.compute.manager [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Took 7.25 seconds to spawn the instance on the hypervisor. [ 784.152209] env[62569]: DEBUG nova.compute.manager [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 784.153057] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7c4767e-0ede-400c-80ab-61b0182c5dc0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.359567] env[62569]: DEBUG oslo_vmware.api [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250000, 'name': CreateSnapshot_Task, 'duration_secs': 0.662496} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.359979] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Created Snapshot of the VM instance {{(pid=62569) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 784.360904] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f15f1d2-1713-4c0e-a99a-a36835e7328e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.400266] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Releasing lock "refresh_cache-06067572-a9fd-43a4-91cd-383a62f53885" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.400677] env[62569]: DEBUG nova.compute.manager [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Instance network_info: |[{"id": "1ae266f0-188a-45de-9a09-d9e21b48bff3", "address": "fa:16:3e:6b:d1:2a", "network": {"id": "f2b76a4d-be64-4c9e-8c25-d2bbe42f59ea", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1621748668-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58f52ff5374a4d5d9969aa2f48f01c8f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2907cce-d529-4809-af05-d29397bed211", "external-id": "nsx-vlan-transportzone-427", "segmentation_id": 427, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1ae266f0-18", "ovs_interfaceid": "1ae266f0-188a-45de-9a09-d9e21b48bff3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 784.401244] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6b:d1:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b2907cce-d529-4809-af05-d29397bed211', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1ae266f0-188a-45de-9a09-d9e21b48bff3', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 784.409224] env[62569]: DEBUG oslo.service.loopingcall [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 784.409508] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 784.409763] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d290888d-9879-47d4-95be-37af55d2f9a5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.432692] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 784.432692] env[62569]: value = "task-1250005" [ 784.432692] env[62569]: _type = "Task" [ 784.432692] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.443552] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250005, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.456631] env[62569]: DEBUG nova.compute.manager [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 784.468458] env[62569]: DEBUG oslo_vmware.api [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Task: {'id': task-1250003, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142837} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.469141] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 784.469343] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 784.469521] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 784.469727] env[62569]: INFO nova.compute.manager [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Took 1.14 seconds to destroy the instance on the hypervisor. [ 784.469992] env[62569]: DEBUG oslo.service.loopingcall [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 784.470232] env[62569]: DEBUG nova.compute.manager [-] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 784.470363] env[62569]: DEBUG nova.network.neutron [-] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 784.493277] env[62569]: DEBUG nova.virt.hardware [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 784.493549] env[62569]: DEBUG nova.virt.hardware [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 784.493711] env[62569]: DEBUG nova.virt.hardware [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 784.493895] env[62569]: DEBUG nova.virt.hardware [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 784.494054] env[62569]: DEBUG nova.virt.hardware [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 784.494207] env[62569]: DEBUG nova.virt.hardware [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 784.494636] env[62569]: DEBUG nova.virt.hardware [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 784.494636] env[62569]: DEBUG nova.virt.hardware [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 784.494760] env[62569]: DEBUG nova.virt.hardware [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 784.494891] env[62569]: DEBUG nova.virt.hardware [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 784.495147] env[62569]: DEBUG nova.virt.hardware [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 784.496058] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16ae0898-52fb-402f-83ec-77266818c4e0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.505218] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c427c4a-800e-4cef-97a9-36e4eb0ce7b0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.573244] env[62569]: DEBUG nova.scheduler.client.report [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 784.585239] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250004, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.445514} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.585511] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 37a4c60a-e660-4c91-bcfe-72638a4397b0/37a4c60a-e660-4c91-bcfe-72638a4397b0.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 784.585726] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 784.586136] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9ae377b1-db00-4333-b360-d70d380ced43 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.593176] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for the task: (returnval){ [ 784.593176] env[62569]: value = "task-1250006" [ 784.593176] env[62569]: _type = "Task" [ 784.593176] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.601765] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250006, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.611923] env[62569]: DEBUG oslo_concurrency.lockutils [None req-29a29fa3-d4bd-4321-9687-ef028a16d9a1 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Lock "0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.865s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.670031] env[62569]: INFO nova.compute.manager [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Took 34.10 seconds to build instance. [ 784.785226] env[62569]: DEBUG nova.compute.manager [req-b790f710-8702-4dcc-affa-257c74ede933 req-2deeb307-29f2-464c-b86e-1a32d238cb4e service nova] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Received event network-changed-1ae266f0-188a-45de-9a09-d9e21b48bff3 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 784.785595] env[62569]: DEBUG nova.compute.manager [req-b790f710-8702-4dcc-affa-257c74ede933 req-2deeb307-29f2-464c-b86e-1a32d238cb4e service nova] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Refreshing instance network info cache due to event network-changed-1ae266f0-188a-45de-9a09-d9e21b48bff3. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 784.785967] env[62569]: DEBUG oslo_concurrency.lockutils [req-b790f710-8702-4dcc-affa-257c74ede933 req-2deeb307-29f2-464c-b86e-1a32d238cb4e service nova] Acquiring lock "refresh_cache-06067572-a9fd-43a4-91cd-383a62f53885" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.786215] env[62569]: DEBUG oslo_concurrency.lockutils [req-b790f710-8702-4dcc-affa-257c74ede933 req-2deeb307-29f2-464c-b86e-1a32d238cb4e service nova] Acquired lock "refresh_cache-06067572-a9fd-43a4-91cd-383a62f53885" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.786463] env[62569]: DEBUG nova.network.neutron [req-b790f710-8702-4dcc-affa-257c74ede933 req-2deeb307-29f2-464c-b86e-1a32d238cb4e service nova] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Refreshing network info cache for port 1ae266f0-188a-45de-9a09-d9e21b48bff3 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 784.885574] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Creating linked-clone VM from snapshot {{(pid=62569) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 784.885940] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-bf9b5f9d-c1fd-4b9f-a928-c8c9f4d329af {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.897155] env[62569]: DEBUG oslo_vmware.api [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 784.897155] env[62569]: value = "task-1250007" [ 784.897155] env[62569]: _type = "Task" [ 784.897155] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.907934] env[62569]: DEBUG oslo_vmware.api [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250007, 'name': CloneVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.944178] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250005, 'name': CreateVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.081173] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.644s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.081173] env[62569]: DEBUG nova.compute.manager [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 785.084166] env[62569]: DEBUG oslo_concurrency.lockutils [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 22.997s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.084378] env[62569]: DEBUG nova.objects.instance [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62569) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 785.107112] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250006, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.131574} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.107589] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 785.108242] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98119670-d272-424b-bd91-d006226f01c4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.123903] env[62569]: DEBUG nova.compute.manager [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 785.136565] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 37a4c60a-e660-4c91-bcfe-72638a4397b0/37a4c60a-e660-4c91-bcfe-72638a4397b0.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 785.138043] env[62569]: DEBUG nova.network.neutron [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Successfully updated port: 06154dac-fb91-4e11-a503-34ae0c293331 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 785.144084] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23d11aa1-5e96-4a34-9fd6-1d3f0f8273c5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.167100] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for the task: (returnval){ [ 785.167100] env[62569]: value = "task-1250008" [ 785.167100] env[62569]: _type = "Task" [ 785.167100] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.173337] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d0fc7b26-4693-4bb6-b8df-e60f843e6a5c tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "b6e5eefc-8c06-445b-a3af-9404578b6179" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.946s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.178007] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250008, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.410332] env[62569]: DEBUG oslo_vmware.api [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250007, 'name': CloneVM_Task} progress is 94%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.441048] env[62569]: DEBUG nova.compute.manager [req-82e18566-89a9-43a3-b35f-0203edd4824c req-f1529b44-b820-4e01-8ea8-adb2e5e38d53 service nova] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Received event network-vif-deleted-ac6451ef-1971-4adb-90a6-9de4848d5c8d {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 785.441048] env[62569]: INFO nova.compute.manager [req-82e18566-89a9-43a3-b35f-0203edd4824c req-f1529b44-b820-4e01-8ea8-adb2e5e38d53 service nova] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Neutron deleted interface ac6451ef-1971-4adb-90a6-9de4848d5c8d; detaching it from the instance and deleting it from the info cache [ 785.441467] env[62569]: DEBUG nova.network.neutron [req-82e18566-89a9-43a3-b35f-0203edd4824c req-f1529b44-b820-4e01-8ea8-adb2e5e38d53 service nova] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.449188] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250005, 'name': CreateVM_Task, 'duration_secs': 0.656232} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.449355] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 785.450022] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.450176] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.450511] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 785.451483] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4326ffe-a2f2-4240-9dd9-dbead7cf6617 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.457070] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for the task: (returnval){ [ 785.457070] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5271d3ed-15fa-6523-7d17-930ec1e8a6ec" [ 785.457070] env[62569]: _type = "Task" [ 785.457070] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.473397] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5271d3ed-15fa-6523-7d17-930ec1e8a6ec, 'name': SearchDatastore_Task, 'duration_secs': 0.009627} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.473742] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.473981] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 785.474412] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.474586] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.474770] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 785.475041] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7f7a0198-814e-425a-8f85-b7f72510cc06 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.482876] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 785.483076] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 785.483775] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc7cd890-d979-4038-9e6a-ae37c6d91251 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.489569] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for the task: (returnval){ [ 785.489569] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ea1170-df17-b001-25fe-f3fef78545d5" [ 785.489569] env[62569]: _type = "Task" [ 785.489569] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.497572] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ea1170-df17-b001-25fe-f3fef78545d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.534970] env[62569]: DEBUG nova.network.neutron [req-b790f710-8702-4dcc-affa-257c74ede933 req-2deeb307-29f2-464c-b86e-1a32d238cb4e service nova] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Updated VIF entry in instance network info cache for port 1ae266f0-188a-45de-9a09-d9e21b48bff3. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 785.535313] env[62569]: DEBUG nova.network.neutron [req-b790f710-8702-4dcc-affa-257c74ede933 req-2deeb307-29f2-464c-b86e-1a32d238cb4e service nova] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Updating instance_info_cache with network_info: [{"id": "1ae266f0-188a-45de-9a09-d9e21b48bff3", "address": "fa:16:3e:6b:d1:2a", "network": {"id": "f2b76a4d-be64-4c9e-8c25-d2bbe42f59ea", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1621748668-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58f52ff5374a4d5d9969aa2f48f01c8f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2907cce-d529-4809-af05-d29397bed211", "external-id": "nsx-vlan-transportzone-427", "segmentation_id": 427, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1ae266f0-18", "ovs_interfaceid": "1ae266f0-188a-45de-9a09-d9e21b48bff3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.556479] env[62569]: DEBUG nova.network.neutron [-] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.589609] env[62569]: DEBUG nova.compute.utils [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 785.591715] env[62569]: DEBUG nova.compute.manager [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 785.595020] env[62569]: DEBUG nova.network.neutron [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 785.648455] env[62569]: DEBUG nova.policy [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f6f94ef6125e4f799800d37471b9eb6f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8c69ac971dbd419f870a9fa68850cb1c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 785.660868] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Acquiring lock "refresh_cache-b3ccc28b-68d0-461d-a67d-b5ad179a80f9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.661018] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Acquired lock "refresh_cache-b3ccc28b-68d0-461d-a67d-b5ad179a80f9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.661174] env[62569]: DEBUG nova.network.neutron [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 785.667965] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.678334] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250008, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.679784] env[62569]: DEBUG nova.compute.manager [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 785.910951] env[62569]: DEBUG oslo_vmware.api [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250007, 'name': CloneVM_Task} progress is 94%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.946881] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9c6c0376-a619-43e9-a77f-acffda8e181d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.957853] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f9f7a6d-ac00-474c-ac26-962a35d53596 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.989181] env[62569]: DEBUG nova.compute.manager [req-82e18566-89a9-43a3-b35f-0203edd4824c req-f1529b44-b820-4e01-8ea8-adb2e5e38d53 service nova] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Detach interface failed, port_id=ac6451ef-1971-4adb-90a6-9de4848d5c8d, reason: Instance 4feb7b78-9f7c-4e64-b0a7-870ed73adf97 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 786.004831] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ea1170-df17-b001-25fe-f3fef78545d5, 'name': SearchDatastore_Task, 'duration_secs': 0.008619} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.005767] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae7cc7d7-deee-4a40-88f1-7c2959c882e0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.011675] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for the task: (returnval){ [ 786.011675] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52acbd86-4f41-f4b4-27f7-c8dccd2dbe2f" [ 786.011675] env[62569]: _type = "Task" [ 786.011675] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.021649] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52acbd86-4f41-f4b4-27f7-c8dccd2dbe2f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.027193] env[62569]: DEBUG nova.network.neutron [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Successfully created port: 303f46d4-6e0a-418c-b62b-17323ab645b6 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 786.037868] env[62569]: DEBUG oslo_concurrency.lockutils [req-b790f710-8702-4dcc-affa-257c74ede933 req-2deeb307-29f2-464c-b86e-1a32d238cb4e service nova] Releasing lock "refresh_cache-06067572-a9fd-43a4-91cd-383a62f53885" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.057847] env[62569]: INFO nova.compute.manager [-] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Took 1.59 seconds to deallocate network for instance. [ 786.098485] env[62569]: DEBUG nova.compute.manager [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 786.101916] env[62569]: DEBUG oslo_concurrency.lockutils [None req-051ccd0e-a1a6-4e22-8c53-034699f9abdc tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.103075] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.690s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.104429] env[62569]: INFO nova.compute.claims [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 786.187082] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250008, 'name': ReconfigVM_Task, 'duration_secs': 0.688376} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.189674] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 37a4c60a-e660-4c91-bcfe-72638a4397b0/37a4c60a-e660-4c91-bcfe-72638a4397b0.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 786.190583] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e8550bdc-cb52-467c-8042-f97b49d68ae5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.197829] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for the task: (returnval){ [ 786.197829] env[62569]: value = "task-1250009" [ 786.197829] env[62569]: _type = "Task" [ 786.197829] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.204017] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.209495] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250009, 'name': Rename_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.218990] env[62569]: DEBUG nova.network.neutron [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 786.411716] env[62569]: DEBUG oslo_vmware.api [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250007, 'name': CloneVM_Task} progress is 95%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.505184] env[62569]: DEBUG nova.network.neutron [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Updating instance_info_cache with network_info: [{"id": "06154dac-fb91-4e11-a503-34ae0c293331", "address": "fa:16:3e:ab:d8:01", "network": {"id": "9c4e9b79-be5c-4b09-b717-0f2b7423bc05", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1323416416-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7633be3f1b38470dae7048754b408f0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d69a4b11-8d65-435f-94a5-28f74a39a718", "external-id": "cl2-zone-59", "segmentation_id": 59, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06154dac-fb", "ovs_interfaceid": "06154dac-fb91-4e11-a503-34ae0c293331", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.524598] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52acbd86-4f41-f4b4-27f7-c8dccd2dbe2f, 'name': SearchDatastore_Task, 'duration_secs': 0.009652} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.525323] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.525638] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 06067572-a9fd-43a4-91cd-383a62f53885/06067572-a9fd-43a4-91cd-383a62f53885.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 786.526905] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ee1b0559-2876-423f-bcf5-8ae059e5fa23 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.534082] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for the task: (returnval){ [ 786.534082] env[62569]: value = "task-1250010" [ 786.534082] env[62569]: _type = "Task" [ 786.534082] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.545266] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250010, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.565493] env[62569]: DEBUG oslo_concurrency.lockutils [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.637726] env[62569]: DEBUG oslo_concurrency.lockutils [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "c9568fcb-745f-4c04-8487-a38a617aab7c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.638069] env[62569]: DEBUG oslo_concurrency.lockutils [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "c9568fcb-745f-4c04-8487-a38a617aab7c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.712884] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250009, 'name': Rename_Task, 'duration_secs': 0.184686} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.713719] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 786.714109] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1f77300b-2f80-4034-a2f9-4b16cc34099b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.726602] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for the task: (returnval){ [ 786.726602] env[62569]: value = "task-1250011" [ 786.726602] env[62569]: _type = "Task" [ 786.726602] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.738561] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250011, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.911545] env[62569]: DEBUG oslo_vmware.api [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250007, 'name': CloneVM_Task, 'duration_secs': 1.680563} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.912731] env[62569]: INFO nova.virt.vmwareapi.vmops [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Created linked-clone VM from snapshot [ 786.912824] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d496c0b8-893f-4a09-8735-28d1e0d99add {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.923674] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Uploading image e370b729-2673-4bfd-b92d-2a4671f01d25 {{(pid=62569) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 786.959840] env[62569]: DEBUG oslo_vmware.rw_handles [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 786.959840] env[62569]: value = "vm-269391" [ 786.959840] env[62569]: _type = "VirtualMachine" [ 786.959840] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 786.960253] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f2b20a74-d78b-4758-a62a-d3c5df9282fc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.970607] env[62569]: DEBUG oslo_vmware.rw_handles [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lease: (returnval){ [ 786.970607] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527152ce-c672-e081-4e31-4cca21b9778b" [ 786.970607] env[62569]: _type = "HttpNfcLease" [ 786.970607] env[62569]: } obtained for exporting VM: (result){ [ 786.970607] env[62569]: value = "vm-269391" [ 786.970607] env[62569]: _type = "VirtualMachine" [ 786.970607] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 786.971050] env[62569]: DEBUG oslo_vmware.api [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the lease: (returnval){ [ 786.971050] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527152ce-c672-e081-4e31-4cca21b9778b" [ 786.971050] env[62569]: _type = "HttpNfcLease" [ 786.971050] env[62569]: } to be ready. {{(pid=62569) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 786.982078] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 786.982078] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527152ce-c672-e081-4e31-4cca21b9778b" [ 786.982078] env[62569]: _type = "HttpNfcLease" [ 786.982078] env[62569]: } is initializing. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 787.011770] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Releasing lock "refresh_cache-b3ccc28b-68d0-461d-a67d-b5ad179a80f9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.012179] env[62569]: DEBUG nova.compute.manager [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Instance network_info: |[{"id": "06154dac-fb91-4e11-a503-34ae0c293331", "address": "fa:16:3e:ab:d8:01", "network": {"id": "9c4e9b79-be5c-4b09-b717-0f2b7423bc05", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1323416416-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7633be3f1b38470dae7048754b408f0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d69a4b11-8d65-435f-94a5-28f74a39a718", "external-id": "cl2-zone-59", "segmentation_id": 59, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06154dac-fb", "ovs_interfaceid": "06154dac-fb91-4e11-a503-34ae0c293331", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 787.012656] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ab:d8:01', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd69a4b11-8d65-435f-94a5-28f74a39a718', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '06154dac-fb91-4e11-a503-34ae0c293331', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 787.025129] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Creating folder: Project (7633be3f1b38470dae7048754b408f0c). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 787.025500] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e20cffa7-e5f4-4424-be51-81d19d833c3e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.036046] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Created folder: Project (7633be3f1b38470dae7048754b408f0c) in parent group-v269330. [ 787.036200] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Creating folder: Instances. Parent ref: group-v269392. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 787.043357] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1c0fffbc-1083-4269-a094-e10fabe874df {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.048719] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250010, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48431} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.048986] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 06067572-a9fd-43a4-91cd-383a62f53885/06067572-a9fd-43a4-91cd-383a62f53885.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 787.049613] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 787.049613] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7ea5ddd3-6eb5-4bd6-8c4f-9a62459690c6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.052460] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Created folder: Instances in parent group-v269392. [ 787.052681] env[62569]: DEBUG oslo.service.loopingcall [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 787.053796] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 787.053796] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0f57ae4c-2abb-474b-82cd-2e8af42535fb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.071609] env[62569]: DEBUG nova.compute.manager [req-79c1cb87-ee8d-403c-9a5b-89136a680014 req-fbcff447-8325-4466-8444-e85b734fc470 service nova] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Received event network-vif-plugged-06154dac-fb91-4e11-a503-34ae0c293331 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 787.071609] env[62569]: DEBUG oslo_concurrency.lockutils [req-79c1cb87-ee8d-403c-9a5b-89136a680014 req-fbcff447-8325-4466-8444-e85b734fc470 service nova] Acquiring lock "b3ccc28b-68d0-461d-a67d-b5ad179a80f9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.071943] env[62569]: DEBUG oslo_concurrency.lockutils [req-79c1cb87-ee8d-403c-9a5b-89136a680014 req-fbcff447-8325-4466-8444-e85b734fc470 service nova] Lock "b3ccc28b-68d0-461d-a67d-b5ad179a80f9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.072034] env[62569]: DEBUG oslo_concurrency.lockutils [req-79c1cb87-ee8d-403c-9a5b-89136a680014 req-fbcff447-8325-4466-8444-e85b734fc470 service nova] Lock "b3ccc28b-68d0-461d-a67d-b5ad179a80f9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.072261] env[62569]: DEBUG nova.compute.manager [req-79c1cb87-ee8d-403c-9a5b-89136a680014 req-fbcff447-8325-4466-8444-e85b734fc470 service nova] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] No waiting events found dispatching network-vif-plugged-06154dac-fb91-4e11-a503-34ae0c293331 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 787.072496] env[62569]: WARNING nova.compute.manager [req-79c1cb87-ee8d-403c-9a5b-89136a680014 req-fbcff447-8325-4466-8444-e85b734fc470 service nova] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Received unexpected event network-vif-plugged-06154dac-fb91-4e11-a503-34ae0c293331 for instance with vm_state building and task_state spawning. [ 787.072653] env[62569]: DEBUG nova.compute.manager [req-79c1cb87-ee8d-403c-9a5b-89136a680014 req-fbcff447-8325-4466-8444-e85b734fc470 service nova] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Received event network-changed-06154dac-fb91-4e11-a503-34ae0c293331 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 787.072831] env[62569]: DEBUG nova.compute.manager [req-79c1cb87-ee8d-403c-9a5b-89136a680014 req-fbcff447-8325-4466-8444-e85b734fc470 service nova] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Refreshing instance network info cache due to event network-changed-06154dac-fb91-4e11-a503-34ae0c293331. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 787.073082] env[62569]: DEBUG oslo_concurrency.lockutils [req-79c1cb87-ee8d-403c-9a5b-89136a680014 req-fbcff447-8325-4466-8444-e85b734fc470 service nova] Acquiring lock "refresh_cache-b3ccc28b-68d0-461d-a67d-b5ad179a80f9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.073244] env[62569]: DEBUG oslo_concurrency.lockutils [req-79c1cb87-ee8d-403c-9a5b-89136a680014 req-fbcff447-8325-4466-8444-e85b734fc470 service nova] Acquired lock "refresh_cache-b3ccc28b-68d0-461d-a67d-b5ad179a80f9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.073419] env[62569]: DEBUG nova.network.neutron [req-79c1cb87-ee8d-403c-9a5b-89136a680014 req-fbcff447-8325-4466-8444-e85b734fc470 service nova] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Refreshing network info cache for port 06154dac-fb91-4e11-a503-34ae0c293331 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 787.077269] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for the task: (returnval){ [ 787.077269] env[62569]: value = "task-1250015" [ 787.077269] env[62569]: _type = "Task" [ 787.077269] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.083686] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 787.083686] env[62569]: value = "task-1250016" [ 787.083686] env[62569]: _type = "Task" [ 787.083686] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.090968] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250015, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.097423] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250016, 'name': CreateVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.113182] env[62569]: DEBUG nova.compute.manager [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 787.122660] env[62569]: DEBUG oslo_concurrency.lockutils [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Acquiring lock "0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.122660] env[62569]: DEBUG oslo_concurrency.lockutils [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Lock "0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.122660] env[62569]: DEBUG oslo_concurrency.lockutils [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Acquiring lock "0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.122660] env[62569]: DEBUG oslo_concurrency.lockutils [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Lock "0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.122924] env[62569]: DEBUG oslo_concurrency.lockutils [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Lock "0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.125980] env[62569]: INFO nova.compute.manager [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Terminating instance [ 787.144555] env[62569]: DEBUG nova.virt.hardware [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 787.144857] env[62569]: DEBUG nova.virt.hardware [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 787.145044] env[62569]: DEBUG nova.virt.hardware [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 787.145238] env[62569]: DEBUG nova.virt.hardware [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 787.145391] env[62569]: DEBUG nova.virt.hardware [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 787.145538] env[62569]: DEBUG nova.virt.hardware [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 787.145762] env[62569]: DEBUG nova.virt.hardware [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 787.145923] env[62569]: DEBUG nova.virt.hardware [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 787.146117] env[62569]: DEBUG nova.virt.hardware [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 787.146371] env[62569]: DEBUG nova.virt.hardware [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 787.146446] env[62569]: DEBUG nova.virt.hardware [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 787.147764] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-616b8b73-0ee3-4a7f-9aef-3b0d660c9650 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.161611] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b294eec4-74cc-4ea8-881f-46c9861a2f40 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.236841] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250011, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.264653] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Acquiring lock "a5b3c345-149b-4dc3-84ff-fe72183efadb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.264938] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Lock "a5b3c345-149b-4dc3-84ff-fe72183efadb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.479491] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 787.479491] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527152ce-c672-e081-4e31-4cca21b9778b" [ 787.479491] env[62569]: _type = "HttpNfcLease" [ 787.479491] env[62569]: } is ready. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 787.480712] env[62569]: DEBUG oslo_vmware.rw_handles [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 787.480712] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527152ce-c672-e081-4e31-4cca21b9778b" [ 787.480712] env[62569]: _type = "HttpNfcLease" [ 787.480712] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 787.481470] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3162d1d9-c139-47a6-959b-79b0e1e9aaa6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.484555] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c43e45ae-a4a7-440d-be1e-93d32643871b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.494938] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cd5ff8b-19d7-4d53-a850-515c3f2ac895 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.498419] env[62569]: DEBUG oslo_vmware.rw_handles [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5249d8dd-86c7-d2ae-a6e1-36b2dd94a3bd/disk-0.vmdk from lease info. {{(pid=62569) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 787.498611] env[62569]: DEBUG oslo_vmware.rw_handles [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5249d8dd-86c7-d2ae-a6e1-36b2dd94a3bd/disk-0.vmdk for reading. {{(pid=62569) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 787.591309] env[62569]: DEBUG nova.network.neutron [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Successfully updated port: 303f46d4-6e0a-418c-b62b-17323ab645b6 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 787.602488] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f572b5a6-cbcd-400b-b4aa-127761ed4069 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.611365] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250015, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.616577] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250016, 'name': CreateVM_Task, 'duration_secs': 0.358066} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.617078] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 787.618454] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4518e5f7-d2a1-4d15-8926-80b4eea4e3a6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.622554] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.622725] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.623076] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 787.623328] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35d3101f-afaa-4623-9cd0-91bc2b379e77 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.634774] env[62569]: DEBUG nova.compute.manager [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 787.635101] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 787.635511] env[62569]: DEBUG nova.compute.provider_tree [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.638752] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-473f62df-e084-45c8-aab8-178c77d25b27 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.641489] env[62569]: DEBUG oslo_vmware.api [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Waiting for the task: (returnval){ [ 787.641489] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5297964a-aff7-7a03-774d-799aaf72b47f" [ 787.641489] env[62569]: _type = "Task" [ 787.641489] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.651317] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 787.651909] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e97e8032-a736-43a6-a69a-10cad6870e41 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.660896] env[62569]: DEBUG oslo_vmware.api [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5297964a-aff7-7a03-774d-799aaf72b47f, 'name': SearchDatastore_Task, 'duration_secs': 0.009377} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.664763] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.664763] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 787.664763] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.664763] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.665389] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 787.665389] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eca6043b-83e6-4f6e-8bc2-78c51a4fc978 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.666111] env[62569]: DEBUG oslo_vmware.api [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Waiting for the task: (returnval){ [ 787.666111] env[62569]: value = "task-1250017" [ 787.666111] env[62569]: _type = "Task" [ 787.666111] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.671149] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 787.671327] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 787.674605] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a84317b-9344-4148-96c6-be4b26d1307a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.677229] env[62569]: DEBUG oslo_vmware.api [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Task: {'id': task-1250017, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.680195] env[62569]: DEBUG oslo_vmware.api [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Waiting for the task: (returnval){ [ 787.680195] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529bc7ae-d6ab-d993-ae82-c3db43773d73" [ 787.680195] env[62569]: _type = "Task" [ 787.680195] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.688410] env[62569]: DEBUG oslo_vmware.api [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529bc7ae-d6ab-d993-ae82-c3db43773d73, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.742382] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250011, 'name': PowerOnVM_Task, 'duration_secs': 0.804305} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.742665] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 787.742876] env[62569]: INFO nova.compute.manager [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Took 8.27 seconds to spawn the instance on the hypervisor. [ 787.743064] env[62569]: DEBUG nova.compute.manager [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 787.745882] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e5c7bd2-af31-441a-a75a-20df5f4a8c16 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.748965] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-307cffe0-2c9e-44ec-bc09-955ebe18bcf5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.858089] env[62569]: DEBUG nova.network.neutron [req-79c1cb87-ee8d-403c-9a5b-89136a680014 req-fbcff447-8325-4466-8444-e85b734fc470 service nova] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Updated VIF entry in instance network info cache for port 06154dac-fb91-4e11-a503-34ae0c293331. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 787.858601] env[62569]: DEBUG nova.network.neutron [req-79c1cb87-ee8d-403c-9a5b-89136a680014 req-fbcff447-8325-4466-8444-e85b734fc470 service nova] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Updating instance_info_cache with network_info: [{"id": "06154dac-fb91-4e11-a503-34ae0c293331", "address": "fa:16:3e:ab:d8:01", "network": {"id": "9c4e9b79-be5c-4b09-b717-0f2b7423bc05", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1323416416-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7633be3f1b38470dae7048754b408f0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d69a4b11-8d65-435f-94a5-28f74a39a718", "external-id": "cl2-zone-59", "segmentation_id": 59, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06154dac-fb", "ovs_interfaceid": "06154dac-fb91-4e11-a503-34ae0c293331", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.095215] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquiring lock "refresh_cache-337d88a2-30b9-4846-929e-042bd7a64a65" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.095563] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquired lock "refresh_cache-337d88a2-30b9-4846-929e-042bd7a64a65" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.095796] env[62569]: DEBUG nova.network.neutron [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 788.102043] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250015, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.138377] env[62569]: DEBUG nova.scheduler.client.report [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 788.176597] env[62569]: DEBUG oslo_vmware.api [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Task: {'id': task-1250017, 'name': PowerOffVM_Task, 'duration_secs': 0.424387} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.176992] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 788.177257] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 788.177612] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2102bec9-1045-4165-a016-e1ea19102ce0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.189865] env[62569]: DEBUG oslo_vmware.api [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529bc7ae-d6ab-d993-ae82-c3db43773d73, 'name': SearchDatastore_Task, 'duration_secs': 0.011013} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.190803] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a28753a4-5deb-4d6e-bb91-497bed809952 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.196042] env[62569]: DEBUG oslo_vmware.api [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Waiting for the task: (returnval){ [ 788.196042] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52cd34f0-c933-a488-b571-de2c51cf5334" [ 788.196042] env[62569]: _type = "Task" [ 788.196042] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.205617] env[62569]: DEBUG oslo_vmware.api [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52cd34f0-c933-a488-b571-de2c51cf5334, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.266829] env[62569]: INFO nova.compute.manager [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Took 37.67 seconds to build instance. [ 788.300776] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 788.302040] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 788.302040] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Deleting the datastore file [datastore2] 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 788.302157] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8dd46423-2159-4091-85f3-fdee6f2477e5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.308923] env[62569]: DEBUG oslo_vmware.api [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Waiting for the task: (returnval){ [ 788.308923] env[62569]: value = "task-1250019" [ 788.308923] env[62569]: _type = "Task" [ 788.308923] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.319628] env[62569]: DEBUG oslo_vmware.api [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Task: {'id': task-1250019, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.361700] env[62569]: DEBUG oslo_concurrency.lockutils [req-79c1cb87-ee8d-403c-9a5b-89136a680014 req-fbcff447-8325-4466-8444-e85b734fc470 service nova] Releasing lock "refresh_cache-b3ccc28b-68d0-461d-a67d-b5ad179a80f9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.603312] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250015, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.646568] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.540s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.646568] env[62569]: DEBUG nova.compute.manager [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 788.648214] env[62569]: DEBUG oslo_concurrency.lockutils [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.772s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.648648] env[62569]: DEBUG nova.objects.instance [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Lazy-loading 'resources' on Instance uuid fd763e12-5b29-48f8-8256-cb6205e0d119 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 788.686667] env[62569]: DEBUG nova.network.neutron [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 788.711721] env[62569]: DEBUG oslo_vmware.api [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52cd34f0-c933-a488-b571-de2c51cf5334, 'name': SearchDatastore_Task, 'duration_secs': 0.010654} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.714474] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.715026] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] b3ccc28b-68d0-461d-a67d-b5ad179a80f9/b3ccc28b-68d0-461d-a67d-b5ad179a80f9.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 788.715780] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ebb44a09-1a55-431d-83be-2bff0837f5f4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.723370] env[62569]: DEBUG oslo_vmware.api [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Waiting for the task: (returnval){ [ 788.723370] env[62569]: value = "task-1250020" [ 788.723370] env[62569]: _type = "Task" [ 788.723370] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.732143] env[62569]: DEBUG oslo_vmware.api [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Task: {'id': task-1250020, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.770838] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "37a4c60a-e660-4c91-bcfe-72638a4397b0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.265s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.800422] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bec307ea-8c18-4ffc-92e1-ba2a0f707041 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "a0b406c3-9466-41bd-9de1-e675cab2ceef" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.800726] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bec307ea-8c18-4ffc-92e1-ba2a0f707041 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "a0b406c3-9466-41bd-9de1-e675cab2ceef" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.820742] env[62569]: DEBUG oslo_vmware.api [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Task: {'id': task-1250019, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146095} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.821230] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 788.821477] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 788.821839] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 788.822236] env[62569]: INFO nova.compute.manager [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Took 1.19 seconds to destroy the instance on the hypervisor. [ 788.822601] env[62569]: DEBUG oslo.service.loopingcall [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 788.822866] env[62569]: DEBUG nova.compute.manager [-] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 788.823073] env[62569]: DEBUG nova.network.neutron [-] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 788.876619] env[62569]: DEBUG nova.network.neutron [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Updating instance_info_cache with network_info: [{"id": "303f46d4-6e0a-418c-b62b-17323ab645b6", "address": "fa:16:3e:68:ca:b8", "network": {"id": "497ea2f3-6685-4d07-9952-b5ccf3c0b2f8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-561818925-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c69ac971dbd419f870a9fa68850cb1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7654928b-7afe-42e3-a18d-68ecc775cefe", "external-id": "cl2-zone-807", "segmentation_id": 807, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap303f46d4-6e", "ovs_interfaceid": "303f46d4-6e0a-418c-b62b-17323ab645b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.105029] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250015, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.149953] env[62569]: DEBUG nova.compute.utils [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 789.151747] env[62569]: DEBUG nova.compute.manager [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 789.151907] env[62569]: DEBUG nova.network.neutron [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 789.201981] env[62569]: DEBUG nova.compute.manager [req-2a65f73f-478c-4ffc-afcc-a7306e598d3f req-ee39ec83-1588-4033-8b64-24dd5fd64a7b service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Received event network-vif-plugged-303f46d4-6e0a-418c-b62b-17323ab645b6 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 789.202252] env[62569]: DEBUG oslo_concurrency.lockutils [req-2a65f73f-478c-4ffc-afcc-a7306e598d3f req-ee39ec83-1588-4033-8b64-24dd5fd64a7b service nova] Acquiring lock "337d88a2-30b9-4846-929e-042bd7a64a65-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.202500] env[62569]: DEBUG oslo_concurrency.lockutils [req-2a65f73f-478c-4ffc-afcc-a7306e598d3f req-ee39ec83-1588-4033-8b64-24dd5fd64a7b service nova] Lock "337d88a2-30b9-4846-929e-042bd7a64a65-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.202685] env[62569]: DEBUG oslo_concurrency.lockutils [req-2a65f73f-478c-4ffc-afcc-a7306e598d3f req-ee39ec83-1588-4033-8b64-24dd5fd64a7b service nova] Lock "337d88a2-30b9-4846-929e-042bd7a64a65-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.202905] env[62569]: DEBUG nova.compute.manager [req-2a65f73f-478c-4ffc-afcc-a7306e598d3f req-ee39ec83-1588-4033-8b64-24dd5fd64a7b service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] No waiting events found dispatching network-vif-plugged-303f46d4-6e0a-418c-b62b-17323ab645b6 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 789.203054] env[62569]: WARNING nova.compute.manager [req-2a65f73f-478c-4ffc-afcc-a7306e598d3f req-ee39ec83-1588-4033-8b64-24dd5fd64a7b service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Received unexpected event network-vif-plugged-303f46d4-6e0a-418c-b62b-17323ab645b6 for instance with vm_state building and task_state spawning. [ 789.203240] env[62569]: DEBUG nova.compute.manager [req-2a65f73f-478c-4ffc-afcc-a7306e598d3f req-ee39ec83-1588-4033-8b64-24dd5fd64a7b service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Received event network-changed-303f46d4-6e0a-418c-b62b-17323ab645b6 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 789.203402] env[62569]: DEBUG nova.compute.manager [req-2a65f73f-478c-4ffc-afcc-a7306e598d3f req-ee39ec83-1588-4033-8b64-24dd5fd64a7b service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Refreshing instance network info cache due to event network-changed-303f46d4-6e0a-418c-b62b-17323ab645b6. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 789.203598] env[62569]: DEBUG oslo_concurrency.lockutils [req-2a65f73f-478c-4ffc-afcc-a7306e598d3f req-ee39ec83-1588-4033-8b64-24dd5fd64a7b service nova] Acquiring lock "refresh_cache-337d88a2-30b9-4846-929e-042bd7a64a65" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.205082] env[62569]: DEBUG nova.policy [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1418d7ccc2a4324aeb586f89c2f817b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4ba7e96d226942bd99ab29fe703c6fea', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 789.233406] env[62569]: DEBUG oslo_vmware.api [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Task: {'id': task-1250020, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.439731} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.235999] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] b3ccc28b-68d0-461d-a67d-b5ad179a80f9/b3ccc28b-68d0-461d-a67d-b5ad179a80f9.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 789.236248] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 789.236898] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9a4f53eb-d352-4ba9-a739-d70d2430ee5b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.243313] env[62569]: DEBUG oslo_vmware.api [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Waiting for the task: (returnval){ [ 789.243313] env[62569]: value = "task-1250021" [ 789.243313] env[62569]: _type = "Task" [ 789.243313] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.253154] env[62569]: DEBUG oslo_vmware.api [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Task: {'id': task-1250021, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.276229] env[62569]: DEBUG nova.compute.manager [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 789.304365] env[62569]: DEBUG nova.compute.utils [None req-bec307ea-8c18-4ffc-92e1-ba2a0f707041 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 789.380386] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Releasing lock "refresh_cache-337d88a2-30b9-4846-929e-042bd7a64a65" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.380784] env[62569]: DEBUG nova.compute.manager [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Instance network_info: |[{"id": "303f46d4-6e0a-418c-b62b-17323ab645b6", "address": "fa:16:3e:68:ca:b8", "network": {"id": "497ea2f3-6685-4d07-9952-b5ccf3c0b2f8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-561818925-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c69ac971dbd419f870a9fa68850cb1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7654928b-7afe-42e3-a18d-68ecc775cefe", "external-id": "cl2-zone-807", "segmentation_id": 807, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap303f46d4-6e", "ovs_interfaceid": "303f46d4-6e0a-418c-b62b-17323ab645b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 789.381118] env[62569]: DEBUG oslo_concurrency.lockutils [req-2a65f73f-478c-4ffc-afcc-a7306e598d3f req-ee39ec83-1588-4033-8b64-24dd5fd64a7b service nova] Acquired lock "refresh_cache-337d88a2-30b9-4846-929e-042bd7a64a65" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.381313] env[62569]: DEBUG nova.network.neutron [req-2a65f73f-478c-4ffc-afcc-a7306e598d3f req-ee39ec83-1588-4033-8b64-24dd5fd64a7b service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Refreshing network info cache for port 303f46d4-6e0a-418c-b62b-17323ab645b6 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 789.382583] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:68:ca:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7654928b-7afe-42e3-a18d-68ecc775cefe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '303f46d4-6e0a-418c-b62b-17323ab645b6', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 789.390420] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Creating folder: Project (8c69ac971dbd419f870a9fa68850cb1c). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 789.395980] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d1d4ec3b-e578-473e-97e2-b93744f2b9e0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.407213] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Created folder: Project (8c69ac971dbd419f870a9fa68850cb1c) in parent group-v269330. [ 789.408028] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Creating folder: Instances. Parent ref: group-v269395. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 789.408028] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0f6ac44d-5e0d-4200-96f9-778ce6eac19c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.417604] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Created folder: Instances in parent group-v269395. [ 789.417883] env[62569]: DEBUG oslo.service.loopingcall [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 789.418121] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 789.418421] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-830e1134-f1f5-4acb-b259-b1a7d4931bbd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.445916] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 789.445916] env[62569]: value = "task-1250024" [ 789.445916] env[62569]: _type = "Task" [ 789.445916] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.454976] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250024, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.487695] env[62569]: DEBUG nova.network.neutron [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Successfully created port: abd9f1dd-a26a-4f38-bee3-041710cbd705 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 789.559109] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4f6e546-823d-4592-ac04-4d5944906353 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.567428] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-914b0949-57e4-4fa6-aceb-3bfe3e615c22 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.604198] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20eed0f9-7f66-42f0-81f1-612d25e98a1b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.616077] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a3b3523-fa1d-4f36-a1e7-50cf27042e6a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.619993] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250015, 'name': ExtendVirtualDisk_Task, 'duration_secs': 2.281779} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.620359] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 789.621521] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5140cc97-4ba3-43cb-8a2e-4e06bf194f67 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.631967] env[62569]: DEBUG nova.network.neutron [-] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.633546] env[62569]: DEBUG nova.compute.provider_tree [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 789.655976] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] 06067572-a9fd-43a4-91cd-383a62f53885/06067572-a9fd-43a4-91cd-383a62f53885.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 789.656693] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5344c56f-af89-4e52-9c64-10e48079511b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.672356] env[62569]: DEBUG nova.compute.manager [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 789.684396] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for the task: (returnval){ [ 789.684396] env[62569]: value = "task-1250025" [ 789.684396] env[62569]: _type = "Task" [ 789.684396] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.693796] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250025, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.723050] env[62569]: DEBUG nova.network.neutron [req-2a65f73f-478c-4ffc-afcc-a7306e598d3f req-ee39ec83-1588-4033-8b64-24dd5fd64a7b service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Updated VIF entry in instance network info cache for port 303f46d4-6e0a-418c-b62b-17323ab645b6. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 789.723398] env[62569]: DEBUG nova.network.neutron [req-2a65f73f-478c-4ffc-afcc-a7306e598d3f req-ee39ec83-1588-4033-8b64-24dd5fd64a7b service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Updating instance_info_cache with network_info: [{"id": "303f46d4-6e0a-418c-b62b-17323ab645b6", "address": "fa:16:3e:68:ca:b8", "network": {"id": "497ea2f3-6685-4d07-9952-b5ccf3c0b2f8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-561818925-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c69ac971dbd419f870a9fa68850cb1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7654928b-7afe-42e3-a18d-68ecc775cefe", "external-id": "cl2-zone-807", "segmentation_id": 807, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap303f46d4-6e", "ovs_interfaceid": "303f46d4-6e0a-418c-b62b-17323ab645b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.757496] env[62569]: DEBUG oslo_vmware.api [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Task: {'id': task-1250021, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.095708} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.757678] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 789.758726] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcae6c49-5d22-4679-92d3-363294225dcb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.784333] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] b3ccc28b-68d0-461d-a67d-b5ad179a80f9/b3ccc28b-68d0-461d-a67d-b5ad179a80f9.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 789.787289] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-63a1cd0f-14f4-4121-b59a-d6e09d92f2c9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.807241] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bec307ea-8c18-4ffc-92e1-ba2a0f707041 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "a0b406c3-9466-41bd-9de1-e675cab2ceef" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.812535] env[62569]: DEBUG oslo_vmware.api [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Waiting for the task: (returnval){ [ 789.812535] env[62569]: value = "task-1250026" [ 789.812535] env[62569]: _type = "Task" [ 789.812535] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.820929] env[62569]: DEBUG oslo_vmware.api [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Task: {'id': task-1250026, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.822014] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.956028] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250024, 'name': CreateVM_Task, 'duration_secs': 0.433152} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.956225] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 789.956933] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.958155] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.958155] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 789.958155] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d53fb67e-3625-4565-ade7-a76d18bd8f13 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.962893] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 789.962893] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]524d3f37-ba8d-c885-c97e-f4ed710a6d2d" [ 789.962893] env[62569]: _type = "Task" [ 789.962893] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.970942] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]524d3f37-ba8d-c885-c97e-f4ed710a6d2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.135434] env[62569]: INFO nova.compute.manager [-] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Took 1.31 seconds to deallocate network for instance. [ 790.157768] env[62569]: ERROR nova.scheduler.client.report [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] [req-10bef203-ab9f-4c31-ab47-3f3104e12c63] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID fa06556a-5785-4014-b8bd-bc240a0cf716. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-10bef203-ab9f-4c31-ab47-3f3104e12c63"}]} [ 790.177125] env[62569]: DEBUG nova.scheduler.client.report [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Refreshing inventories for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 790.194566] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250025, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.196873] env[62569]: DEBUG nova.scheduler.client.report [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Updating ProviderTree inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 790.197731] env[62569]: DEBUG nova.compute.provider_tree [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 790.208315] env[62569]: DEBUG nova.scheduler.client.report [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Refreshing aggregate associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, aggregates: None {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 790.225749] env[62569]: DEBUG oslo_concurrency.lockutils [req-2a65f73f-478c-4ffc-afcc-a7306e598d3f req-ee39ec83-1588-4033-8b64-24dd5fd64a7b service nova] Releasing lock "refresh_cache-337d88a2-30b9-4846-929e-042bd7a64a65" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.228491] env[62569]: DEBUG nova.scheduler.client.report [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Refreshing trait associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 790.323693] env[62569]: DEBUG oslo_vmware.api [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Task: {'id': task-1250026, 'name': ReconfigVM_Task, 'duration_secs': 0.271792} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.324094] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Reconfigured VM instance instance-00000040 to attach disk [datastore2] b3ccc28b-68d0-461d-a67d-b5ad179a80f9/b3ccc28b-68d0-461d-a67d-b5ad179a80f9.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 790.325615] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c196f614-4f60-4a9b-a899-32c991a2dcae {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.333504] env[62569]: DEBUG oslo_vmware.api [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Waiting for the task: (returnval){ [ 790.333504] env[62569]: value = "task-1250027" [ 790.333504] env[62569]: _type = "Task" [ 790.333504] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.345656] env[62569]: DEBUG oslo_vmware.api [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Task: {'id': task-1250027, 'name': Rename_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.474666] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]524d3f37-ba8d-c885-c97e-f4ed710a6d2d, 'name': SearchDatastore_Task, 'duration_secs': 0.008858} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.477487] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.477748] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 790.477994] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.478178] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.478364] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 790.478901] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8e086f26-314b-4dc1-9768-9c84bed847a7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.486760] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 790.486980] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 790.487767] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4b7330b-3105-4c31-9cf7-537ec48f9335 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.496432] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 790.496432] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528cbf7a-468a-9a35-9070-12a45972db6c" [ 790.496432] env[62569]: _type = "Task" [ 790.496432] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.504490] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528cbf7a-468a-9a35-9070-12a45972db6c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.580045] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fbc5b06-ac0b-44f2-8b67-b867297ffea3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.588494] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-616a8d7a-f04d-4768-abb1-4a879d67fd28 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.618955] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3907c1f2-ecb8-49b8-9b3e-40c32fa9a136 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.626708] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd29756d-e456-4192-bf22-bed4774f00cb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.640514] env[62569]: DEBUG nova.compute.provider_tree [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 790.645613] env[62569]: DEBUG oslo_concurrency.lockutils [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.685642] env[62569]: DEBUG nova.compute.manager [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 790.696884] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250025, 'name': ReconfigVM_Task, 'duration_secs': 0.513082} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.697198] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Reconfigured VM instance instance-0000003f to attach disk [datastore1] 06067572-a9fd-43a4-91cd-383a62f53885/06067572-a9fd-43a4-91cd-383a62f53885.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 790.697826] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ae50a3c4-ff51-4bc5-bf56-e65d8b371aa3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.704781] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for the task: (returnval){ [ 790.704781] env[62569]: value = "task-1250028" [ 790.704781] env[62569]: _type = "Task" [ 790.704781] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.709788] env[62569]: DEBUG nova.virt.hardware [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 790.709956] env[62569]: DEBUG nova.virt.hardware [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 790.710102] env[62569]: DEBUG nova.virt.hardware [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 790.710284] env[62569]: DEBUG nova.virt.hardware [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 790.710461] env[62569]: DEBUG nova.virt.hardware [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 790.710625] env[62569]: DEBUG nova.virt.hardware [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 790.710838] env[62569]: DEBUG nova.virt.hardware [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 790.711009] env[62569]: DEBUG nova.virt.hardware [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 790.711189] env[62569]: DEBUG nova.virt.hardware [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 790.711355] env[62569]: DEBUG nova.virt.hardware [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 790.711599] env[62569]: DEBUG nova.virt.hardware [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 790.712446] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68add453-81ab-466e-b447-a647e8bf4d3b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.720376] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250028, 'name': Rename_Task} progress is 10%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.723514] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10bdd170-4e54-49e0-acd7-6cb564f1c860 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.844255] env[62569]: DEBUG oslo_vmware.api [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Task: {'id': task-1250027, 'name': Rename_Task, 'duration_secs': 0.1466} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.844568] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 790.844842] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ad08fb80-78d1-48bc-9a7c-99ad1a2cf5e4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.851331] env[62569]: DEBUG oslo_vmware.api [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Waiting for the task: (returnval){ [ 790.851331] env[62569]: value = "task-1250029" [ 790.851331] env[62569]: _type = "Task" [ 790.851331] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.859314] env[62569]: DEBUG oslo_vmware.api [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Task: {'id': task-1250029, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.881274] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bec307ea-8c18-4ffc-92e1-ba2a0f707041 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "a0b406c3-9466-41bd-9de1-e675cab2ceef" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.881563] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bec307ea-8c18-4ffc-92e1-ba2a0f707041 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "a0b406c3-9466-41bd-9de1-e675cab2ceef" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.881803] env[62569]: INFO nova.compute.manager [None req-bec307ea-8c18-4ffc-92e1-ba2a0f707041 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Attaching volume 025e7ad1-7cda-4045-99fb-506305a5e0cf to /dev/sdb [ 790.919741] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffb62404-68f4-4ca7-8e8b-f06664957882 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.927158] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c2afab-e4bd-435e-b654-673202c5d94f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.940957] env[62569]: DEBUG nova.virt.block_device [None req-bec307ea-8c18-4ffc-92e1-ba2a0f707041 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Updating existing volume attachment record: 8da000d3-97bd-45c4-a2f2-242288dce630 {{(pid=62569) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 791.010687] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528cbf7a-468a-9a35-9070-12a45972db6c, 'name': SearchDatastore_Task, 'duration_secs': 0.009178} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.011619] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6398e67a-6a1c-47fb-9a36-ae899cacb529 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.017196] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 791.017196] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]522f2d39-df7c-73ed-77d0-82e8f6556ee7" [ 791.017196] env[62569]: _type = "Task" [ 791.017196] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.025685] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]522f2d39-df7c-73ed-77d0-82e8f6556ee7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.031363] env[62569]: DEBUG nova.network.neutron [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Successfully updated port: abd9f1dd-a26a-4f38-bee3-041710cbd705 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 791.144193] env[62569]: DEBUG nova.scheduler.client.report [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 791.213960] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250028, 'name': Rename_Task, 'duration_secs': 0.347175} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.214295] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 791.214566] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b05a27b2-9a5d-4f29-80a9-6d2f698a2aeb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.221625] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for the task: (returnval){ [ 791.221625] env[62569]: value = "task-1250032" [ 791.221625] env[62569]: _type = "Task" [ 791.221625] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.233285] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250032, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.239675] env[62569]: DEBUG nova.compute.manager [req-f84e7d17-4802-48f9-8efe-d2315021493c req-91b5a940-9dde-43a5-a244-e4a7e2718b37 service nova] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Received event network-vif-deleted-62cedd28-f07d-4ca0-89b4-ffbc290ff45f {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 791.239975] env[62569]: DEBUG nova.compute.manager [req-f84e7d17-4802-48f9-8efe-d2315021493c req-91b5a940-9dde-43a5-a244-e4a7e2718b37 service nova] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Received event network-vif-plugged-abd9f1dd-a26a-4f38-bee3-041710cbd705 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 791.241990] env[62569]: DEBUG oslo_concurrency.lockutils [req-f84e7d17-4802-48f9-8efe-d2315021493c req-91b5a940-9dde-43a5-a244-e4a7e2718b37 service nova] Acquiring lock "eb70341b-4282-4eca-b6a2-374db7c521c5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.242415] env[62569]: DEBUG oslo_concurrency.lockutils [req-f84e7d17-4802-48f9-8efe-d2315021493c req-91b5a940-9dde-43a5-a244-e4a7e2718b37 service nova] Lock "eb70341b-4282-4eca-b6a2-374db7c521c5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.242717] env[62569]: DEBUG oslo_concurrency.lockutils [req-f84e7d17-4802-48f9-8efe-d2315021493c req-91b5a940-9dde-43a5-a244-e4a7e2718b37 service nova] Lock "eb70341b-4282-4eca-b6a2-374db7c521c5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.243046] env[62569]: DEBUG nova.compute.manager [req-f84e7d17-4802-48f9-8efe-d2315021493c req-91b5a940-9dde-43a5-a244-e4a7e2718b37 service nova] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] No waiting events found dispatching network-vif-plugged-abd9f1dd-a26a-4f38-bee3-041710cbd705 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 791.243340] env[62569]: WARNING nova.compute.manager [req-f84e7d17-4802-48f9-8efe-d2315021493c req-91b5a940-9dde-43a5-a244-e4a7e2718b37 service nova] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Received unexpected event network-vif-plugged-abd9f1dd-a26a-4f38-bee3-041710cbd705 for instance with vm_state building and task_state spawning. [ 791.243616] env[62569]: DEBUG nova.compute.manager [req-f84e7d17-4802-48f9-8efe-d2315021493c req-91b5a940-9dde-43a5-a244-e4a7e2718b37 service nova] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Received event network-changed-abd9f1dd-a26a-4f38-bee3-041710cbd705 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 791.243894] env[62569]: DEBUG nova.compute.manager [req-f84e7d17-4802-48f9-8efe-d2315021493c req-91b5a940-9dde-43a5-a244-e4a7e2718b37 service nova] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Refreshing instance network info cache due to event network-changed-abd9f1dd-a26a-4f38-bee3-041710cbd705. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 791.244219] env[62569]: DEBUG oslo_concurrency.lockutils [req-f84e7d17-4802-48f9-8efe-d2315021493c req-91b5a940-9dde-43a5-a244-e4a7e2718b37 service nova] Acquiring lock "refresh_cache-eb70341b-4282-4eca-b6a2-374db7c521c5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.244448] env[62569]: DEBUG oslo_concurrency.lockutils [req-f84e7d17-4802-48f9-8efe-d2315021493c req-91b5a940-9dde-43a5-a244-e4a7e2718b37 service nova] Acquired lock "refresh_cache-eb70341b-4282-4eca-b6a2-374db7c521c5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.244690] env[62569]: DEBUG nova.network.neutron [req-f84e7d17-4802-48f9-8efe-d2315021493c req-91b5a940-9dde-43a5-a244-e4a7e2718b37 service nova] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Refreshing network info cache for port abd9f1dd-a26a-4f38-bee3-041710cbd705 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 791.361831] env[62569]: DEBUG oslo_vmware.api [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Task: {'id': task-1250029, 'name': PowerOnVM_Task, 'duration_secs': 0.47838} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.362168] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 791.362383] env[62569]: INFO nova.compute.manager [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Took 6.91 seconds to spawn the instance on the hypervisor. [ 791.362574] env[62569]: DEBUG nova.compute.manager [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 791.363411] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29806425-7a9e-4695-b3f4-bae803fd024e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.533977] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "refresh_cache-eb70341b-4282-4eca-b6a2-374db7c521c5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.533977] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]522f2d39-df7c-73ed-77d0-82e8f6556ee7, 'name': SearchDatastore_Task, 'duration_secs': 0.01044} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.534513] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.534873] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 337d88a2-30b9-4846-929e-042bd7a64a65/337d88a2-30b9-4846-929e-042bd7a64a65.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 791.535233] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d3c69aa1-3e7c-42d7-b93f-88d55b0f73ac {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.543214] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 791.543214] env[62569]: value = "task-1250034" [ 791.543214] env[62569]: _type = "Task" [ 791.543214] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.552218] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250034, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.649689] env[62569]: DEBUG oslo_concurrency.lockutils [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.652606] env[62569]: DEBUG oslo_concurrency.lockutils [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.351s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.652892] env[62569]: DEBUG nova.objects.instance [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lazy-loading 'resources' on Instance uuid 015140aa-8336-40e2-856c-723277a48f6e {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 791.677089] env[62569]: INFO nova.scheduler.client.report [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Deleted allocations for instance fd763e12-5b29-48f8-8256-cb6205e0d119 [ 791.736440] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250032, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.783131] env[62569]: DEBUG nova.network.neutron [req-f84e7d17-4802-48f9-8efe-d2315021493c req-91b5a940-9dde-43a5-a244-e4a7e2718b37 service nova] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.881443] env[62569]: DEBUG nova.network.neutron [req-f84e7d17-4802-48f9-8efe-d2315021493c req-91b5a940-9dde-43a5-a244-e4a7e2718b37 service nova] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.887793] env[62569]: INFO nova.compute.manager [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Took 37.16 seconds to build instance. [ 792.054301] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250034, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.187247] env[62569]: DEBUG oslo_concurrency.lockutils [None req-079b8282-cc3c-42f8-abb8-69989a08826e tempest-ServersAaction247Test-7521631 tempest-ServersAaction247Test-7521631-project-member] Lock "fd763e12-5b29-48f8-8256-cb6205e0d119" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.539s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.237148] env[62569]: DEBUG oslo_vmware.api [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250032, 'name': PowerOnVM_Task, 'duration_secs': 0.701518} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.237500] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 792.238128] env[62569]: INFO nova.compute.manager [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Took 10.29 seconds to spawn the instance on the hypervisor. [ 792.238879] env[62569]: DEBUG nova.compute.manager [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 792.242418] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdd89b11-d5d8-4c99-8606-27f4f6aa46a1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.388072] env[62569]: DEBUG oslo_concurrency.lockutils [req-f84e7d17-4802-48f9-8efe-d2315021493c req-91b5a940-9dde-43a5-a244-e4a7e2718b37 service nova] Releasing lock "refresh_cache-eb70341b-4282-4eca-b6a2-374db7c521c5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.388556] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired lock "refresh_cache-eb70341b-4282-4eca-b6a2-374db7c521c5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.388630] env[62569]: DEBUG nova.network.neutron [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 792.391870] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ff881997-6fe4-4b28-94d9-635fdda5e430 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Lock "b3ccc28b-68d0-461d-a67d-b5ad179a80f9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.626s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.528522] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b8e7eaf-b11a-455c-9666-e77d34462009 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.537168] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e72b6c2e-45f7-449c-98de-3aa8f239da75 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.572939] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ceb9c70-4b10-42c0-a8d4-06be20b615e3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.578824] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250034, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.637272} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.579514] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 337d88a2-30b9-4846-929e-042bd7a64a65/337d88a2-30b9-4846-929e-042bd7a64a65.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 792.579724] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 792.580012] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4bc7ad29-17de-4f0b-9ce4-083179448315 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.585790] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-592aa7c8-e4ef-4261-878d-6015eef9bb6f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.592353] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 792.592353] env[62569]: value = "task-1250035" [ 792.592353] env[62569]: _type = "Task" [ 792.592353] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.607281] env[62569]: DEBUG nova.compute.provider_tree [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.612520] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250035, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.764915] env[62569]: INFO nova.compute.manager [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Took 41.16 seconds to build instance. [ 792.897203] env[62569]: DEBUG nova.compute.manager [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 792.940990] env[62569]: DEBUG nova.network.neutron [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.098196] env[62569]: DEBUG nova.network.neutron [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Updating instance_info_cache with network_info: [{"id": "abd9f1dd-a26a-4f38-bee3-041710cbd705", "address": "fa:16:3e:15:fa:ec", "network": {"id": "c61b7867-7897-43ae-8f5b-19834c4390ea", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-512066189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ba7e96d226942bd99ab29fe703c6fea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapabd9f1dd-a2", "ovs_interfaceid": "abd9f1dd-a26a-4f38-bee3-041710cbd705", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.106023] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250035, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07119} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.106320] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 793.107468] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2591fa4a-2e12-41aa-8995-ea4f3d8c3566 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.113156] env[62569]: DEBUG nova.scheduler.client.report [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 793.141476] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 337d88a2-30b9-4846-929e-042bd7a64a65/337d88a2-30b9-4846-929e-042bd7a64a65.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 793.142200] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2ae752c-c3a8-41be-8486-5b57210a06d0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.164638] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 793.164638] env[62569]: value = "task-1250036" [ 793.164638] env[62569]: _type = "Task" [ 793.164638] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.176211] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250036, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.269704] env[62569]: DEBUG oslo_concurrency.lockutils [None req-053b1d92-0e3e-49af-9d3f-72825c190ed7 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "06067572-a9fd-43a4-91cd-383a62f53885" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.737s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.427812] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.601353] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Releasing lock "refresh_cache-eb70341b-4282-4eca-b6a2-374db7c521c5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.601790] env[62569]: DEBUG nova.compute.manager [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Instance network_info: |[{"id": "abd9f1dd-a26a-4f38-bee3-041710cbd705", "address": "fa:16:3e:15:fa:ec", "network": {"id": "c61b7867-7897-43ae-8f5b-19834c4390ea", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-512066189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ba7e96d226942bd99ab29fe703c6fea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapabd9f1dd-a2", "ovs_interfaceid": "abd9f1dd-a26a-4f38-bee3-041710cbd705", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 793.602285] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:15:fa:ec', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6685c85e-be1e-4b7b-a6cc-3e50e59b6567', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'abd9f1dd-a26a-4f38-bee3-041710cbd705', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 793.610314] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Creating folder: Project (4ba7e96d226942bd99ab29fe703c6fea). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 793.610712] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f911685f-3faa-4158-acfc-89b83cbeca6f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.621196] env[62569]: DEBUG oslo_concurrency.lockutils [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.968s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.624449] env[62569]: DEBUG oslo_concurrency.lockutils [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.314s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.624619] env[62569]: DEBUG nova.objects.instance [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Lazy-loading 'resources' on Instance uuid c515d85e-bcb5-4bac-bacb-1e558f38171f {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 793.627333] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Created folder: Project (4ba7e96d226942bd99ab29fe703c6fea) in parent group-v269330. [ 793.630672] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Creating folder: Instances. Parent ref: group-v269400. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 793.631369] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-865afa5d-9694-449a-9e54-92a9543d85c9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.647836] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Created folder: Instances in parent group-v269400. [ 793.648217] env[62569]: DEBUG oslo.service.loopingcall [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 793.650010] env[62569]: INFO nova.scheduler.client.report [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Deleted allocations for instance 015140aa-8336-40e2-856c-723277a48f6e [ 793.651289] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 793.654589] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c6e06fbc-389d-4d4e-af39-a58bdf38d562 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.683026] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250036, 'name': ReconfigVM_Task, 'duration_secs': 0.510194} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.684479] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 337d88a2-30b9-4846-929e-042bd7a64a65/337d88a2-30b9-4846-929e-042bd7a64a65.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 793.685198] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 793.685198] env[62569]: value = "task-1250040" [ 793.685198] env[62569]: _type = "Task" [ 793.685198] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.685433] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-910994e7-89c4-4582-8b8c-562eb558e58c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.700213] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250040, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.702331] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 793.702331] env[62569]: value = "task-1250041" [ 793.702331] env[62569]: _type = "Task" [ 793.702331] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.713657] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250041, 'name': Rename_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.776884] env[62569]: DEBUG nova.compute.manager [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 794.179492] env[62569]: DEBUG oslo_concurrency.lockutils [None req-65279cb5-c510-419a-a35b-43c60124441c tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lock "015140aa-8336-40e2-856c-723277a48f6e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.271s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.200538] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250040, 'name': CreateVM_Task, 'duration_secs': 0.366307} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.200776] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 794.201581] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.201823] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.202176] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 794.202687] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2cd93d38-3be7-4d77-9aa6-e93de678ebd5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.215714] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 794.215714] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5272ec52-6230-8118-5580-672c638e433c" [ 794.215714] env[62569]: _type = "Task" [ 794.215714] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.220178] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250041, 'name': Rename_Task, 'duration_secs': 0.209403} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.223922] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 794.227188] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-000c7c1a-d555-4936-9b6d-9ebdeb1d8e1c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.237800] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5272ec52-6230-8118-5580-672c638e433c, 'name': SearchDatastore_Task, 'duration_secs': 0.017877} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.239485] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.239581] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 794.239895] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.240170] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.240513] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 794.240819] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 794.240819] env[62569]: value = "task-1250042" [ 794.240819] env[62569]: _type = "Task" [ 794.240819] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.241047] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a8193f7f-440f-4408-ab27-d53e082e9a41 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.258630] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250042, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.263999] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 794.263999] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 794.267063] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae260354-6beb-40bf-b40a-f270586e6b04 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.273426] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 794.273426] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52788b89-4cee-14bf-325f-f5b6a9de4299" [ 794.273426] env[62569]: _type = "Task" [ 794.273426] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.286427] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52788b89-4cee-14bf-325f-f5b6a9de4299, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.303299] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.403925] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "37a4c60a-e660-4c91-bcfe-72638a4397b0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.404220] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "37a4c60a-e660-4c91-bcfe-72638a4397b0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.404435] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "37a4c60a-e660-4c91-bcfe-72638a4397b0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.404617] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "37a4c60a-e660-4c91-bcfe-72638a4397b0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.404790] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "37a4c60a-e660-4c91-bcfe-72638a4397b0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.410562] env[62569]: INFO nova.compute.manager [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Terminating instance [ 794.516519] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eced803d-dcac-4c5e-9dd3-0572bc53abd6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.525799] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-829d89c2-0c70-42a9-a69d-705ec370b529 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.562168] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63880a66-103b-4b7f-8dff-40cdcb1aa626 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.571730] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e32d6dee-54b0-4a2a-8b4d-ef50b30967de {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.590805] env[62569]: DEBUG nova.compute.provider_tree [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.754973] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250042, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.757407] env[62569]: DEBUG nova.compute.manager [req-345ce5b6-ac05-4a08-a3f4-d5f48613f8e2 req-7aacbc56-7184-4df4-8c22-fd52a4274089 service nova] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Received event network-changed-06154dac-fb91-4e11-a503-34ae0c293331 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 794.757599] env[62569]: DEBUG nova.compute.manager [req-345ce5b6-ac05-4a08-a3f4-d5f48613f8e2 req-7aacbc56-7184-4df4-8c22-fd52a4274089 service nova] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Refreshing instance network info cache due to event network-changed-06154dac-fb91-4e11-a503-34ae0c293331. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 794.757814] env[62569]: DEBUG oslo_concurrency.lockutils [req-345ce5b6-ac05-4a08-a3f4-d5f48613f8e2 req-7aacbc56-7184-4df4-8c22-fd52a4274089 service nova] Acquiring lock "refresh_cache-b3ccc28b-68d0-461d-a67d-b5ad179a80f9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.757958] env[62569]: DEBUG oslo_concurrency.lockutils [req-345ce5b6-ac05-4a08-a3f4-d5f48613f8e2 req-7aacbc56-7184-4df4-8c22-fd52a4274089 service nova] Acquired lock "refresh_cache-b3ccc28b-68d0-461d-a67d-b5ad179a80f9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.758239] env[62569]: DEBUG nova.network.neutron [req-345ce5b6-ac05-4a08-a3f4-d5f48613f8e2 req-7aacbc56-7184-4df4-8c22-fd52a4274089 service nova] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Refreshing network info cache for port 06154dac-fb91-4e11-a503-34ae0c293331 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 794.785737] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52788b89-4cee-14bf-325f-f5b6a9de4299, 'name': SearchDatastore_Task, 'duration_secs': 0.017032} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.786653] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6081d59c-d283-4e01-8ddc-833acd1a18c4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.793883] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 794.793883] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52fa214c-0688-ccd3-0f04-f2a3a7fecaf5" [ 794.793883] env[62569]: _type = "Task" [ 794.793883] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.804115] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52fa214c-0688-ccd3-0f04-f2a3a7fecaf5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.918223] env[62569]: DEBUG nova.compute.manager [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 794.918655] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 794.919734] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12bfed20-ebc8-4074-85c2-2fb273f67910 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.929938] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 794.930433] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bdc3a94b-eb53-4d9e-a209-55ed418c0d66 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.938758] env[62569]: DEBUG oslo_vmware.api [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for the task: (returnval){ [ 794.938758] env[62569]: value = "task-1250043" [ 794.938758] env[62569]: _type = "Task" [ 794.938758] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.949326] env[62569]: DEBUG oslo_vmware.api [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250043, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.093953] env[62569]: DEBUG nova.scheduler.client.report [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 795.174400] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquiring lock "94392d48-223d-4205-9627-25488a468769" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.174936] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lock "94392d48-223d-4205-9627-25488a468769" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.174936] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquiring lock "94392d48-223d-4205-9627-25488a468769-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.175167] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lock "94392d48-223d-4205-9627-25488a468769-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.175356] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lock "94392d48-223d-4205-9627-25488a468769-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.177581] env[62569]: INFO nova.compute.manager [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Terminating instance [ 795.198289] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "06067572-a9fd-43a4-91cd-383a62f53885" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.198890] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "06067572-a9fd-43a4-91cd-383a62f53885" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.199173] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "06067572-a9fd-43a4-91cd-383a62f53885-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.199442] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "06067572-a9fd-43a4-91cd-383a62f53885-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.199688] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "06067572-a9fd-43a4-91cd-383a62f53885-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.201809] env[62569]: INFO nova.compute.manager [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Terminating instance [ 795.259715] env[62569]: DEBUG oslo_vmware.api [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250042, 'name': PowerOnVM_Task, 'duration_secs': 0.794813} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.263158] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 795.263476] env[62569]: INFO nova.compute.manager [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Took 8.15 seconds to spawn the instance on the hypervisor. [ 795.263758] env[62569]: DEBUG nova.compute.manager [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 795.264942] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e0599c9-a441-4eee-9abb-bdb214952243 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.304941] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52fa214c-0688-ccd3-0f04-f2a3a7fecaf5, 'name': SearchDatastore_Task, 'duration_secs': 0.016523} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.306896] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.307193] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] eb70341b-4282-4eca-b6a2-374db7c521c5/eb70341b-4282-4eca-b6a2-374db7c521c5.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 795.307481] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fb5ae12f-df0f-4173-826a-c8bdf0d1da21 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.317221] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 795.317221] env[62569]: value = "task-1250044" [ 795.317221] env[62569]: _type = "Task" [ 795.317221] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.327960] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250044, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.454559] env[62569]: DEBUG oslo_vmware.api [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250043, 'name': PowerOffVM_Task, 'duration_secs': 0.290923} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.454850] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 795.455078] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 795.455761] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c2cb100c-4726-4a06-9f49-6ffebf54821c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.497095] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-bec307ea-8c18-4ffc-92e1-ba2a0f707041 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Volume attach. Driver type: vmdk {{(pid=62569) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 795.497519] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-bec307ea-8c18-4ffc-92e1-ba2a0f707041 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269399', 'volume_id': '025e7ad1-7cda-4045-99fb-506305a5e0cf', 'name': 'volume-025e7ad1-7cda-4045-99fb-506305a5e0cf', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a0b406c3-9466-41bd-9de1-e675cab2ceef', 'attached_at': '', 'detached_at': '', 'volume_id': '025e7ad1-7cda-4045-99fb-506305a5e0cf', 'serial': '025e7ad1-7cda-4045-99fb-506305a5e0cf'} {{(pid=62569) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 795.498699] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00594f60-5b15-4291-a562-ce569bd04671 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.518707] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4786791-6fdc-48b2-92f1-ebc7480011e5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.551105] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-bec307ea-8c18-4ffc-92e1-ba2a0f707041 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] volume-025e7ad1-7cda-4045-99fb-506305a5e0cf/volume-025e7ad1-7cda-4045-99fb-506305a5e0cf.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 795.552279] env[62569]: DEBUG nova.network.neutron [req-345ce5b6-ac05-4a08-a3f4-d5f48613f8e2 req-7aacbc56-7184-4df4-8c22-fd52a4274089 service nova] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Updated VIF entry in instance network info cache for port 06154dac-fb91-4e11-a503-34ae0c293331. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 795.552624] env[62569]: DEBUG nova.network.neutron [req-345ce5b6-ac05-4a08-a3f4-d5f48613f8e2 req-7aacbc56-7184-4df4-8c22-fd52a4274089 service nova] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Updating instance_info_cache with network_info: [{"id": "06154dac-fb91-4e11-a503-34ae0c293331", "address": "fa:16:3e:ab:d8:01", "network": {"id": "9c4e9b79-be5c-4b09-b717-0f2b7423bc05", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1323416416-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7633be3f1b38470dae7048754b408f0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d69a4b11-8d65-435f-94a5-28f74a39a718", "external-id": "cl2-zone-59", "segmentation_id": 59, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06154dac-fb", "ovs_interfaceid": "06154dac-fb91-4e11-a503-34ae0c293331", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.553946] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5a26ddc3-2341-495d-8b57-7a0a6be61050 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.576977] env[62569]: DEBUG oslo_vmware.api [None req-bec307ea-8c18-4ffc-92e1-ba2a0f707041 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 795.576977] env[62569]: value = "task-1250046" [ 795.576977] env[62569]: _type = "Task" [ 795.576977] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.586345] env[62569]: DEBUG oslo_vmware.api [None req-bec307ea-8c18-4ffc-92e1-ba2a0f707041 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250046, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.599798] env[62569]: DEBUG oslo_concurrency.lockutils [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.975s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.602360] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.471s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.603914] env[62569]: INFO nova.compute.claims [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 795.629413] env[62569]: INFO nova.scheduler.client.report [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Deleted allocations for instance c515d85e-bcb5-4bac-bacb-1e558f38171f [ 795.681691] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquiring lock "refresh_cache-94392d48-223d-4205-9627-25488a468769" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.681900] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquired lock "refresh_cache-94392d48-223d-4205-9627-25488a468769" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.682117] env[62569]: DEBUG nova.network.neutron [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 795.706775] env[62569]: DEBUG nova.compute.manager [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 795.706775] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 795.707566] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd145e3a-09f2-43b1-bc85-bc298eafc175 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.716315] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 795.716618] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-69e55639-c481-42b7-bda3-347b05c68dea {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.724384] env[62569]: DEBUG oslo_vmware.api [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for the task: (returnval){ [ 795.724384] env[62569]: value = "task-1250047" [ 795.724384] env[62569]: _type = "Task" [ 795.724384] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.734112] env[62569]: DEBUG oslo_vmware.api [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250047, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.796935] env[62569]: INFO nova.compute.manager [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Took 33.90 seconds to build instance. [ 795.808717] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 795.809113] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 795.809282] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Deleting the datastore file [datastore2] 37a4c60a-e660-4c91-bcfe-72638a4397b0 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 795.809583] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c942418d-e5e9-4459-8580-84fb2cf71ae9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.819091] env[62569]: DEBUG oslo_vmware.api [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for the task: (returnval){ [ 795.819091] env[62569]: value = "task-1250048" [ 795.819091] env[62569]: _type = "Task" [ 795.819091] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.840726] env[62569]: DEBUG oslo_vmware.api [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250048, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.842182] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250044, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.070419] env[62569]: DEBUG oslo_concurrency.lockutils [req-345ce5b6-ac05-4a08-a3f4-d5f48613f8e2 req-7aacbc56-7184-4df4-8c22-fd52a4274089 service nova] Releasing lock "refresh_cache-b3ccc28b-68d0-461d-a67d-b5ad179a80f9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.088391] env[62569]: DEBUG oslo_vmware.api [None req-bec307ea-8c18-4ffc-92e1-ba2a0f707041 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250046, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.139925] env[62569]: DEBUG oslo_concurrency.lockutils [None req-eb51dd07-fbfc-4fb5-a465-791f7147c451 tempest-SecurityGroupsTestJSON-31016293 tempest-SecurityGroupsTestJSON-31016293-project-member] Lock "c515d85e-bcb5-4bac-bacb-1e558f38171f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.509s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.208891] env[62569]: DEBUG nova.network.neutron [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.234684] env[62569]: DEBUG oslo_vmware.api [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250047, 'name': PowerOffVM_Task, 'duration_secs': 0.312005} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.235237] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 796.235587] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 796.235980] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0885b4ec-fed4-4763-b2c1-ba8a103f0085 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.259327] env[62569]: DEBUG nova.network.neutron [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.301224] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3c87a76b-59de-41b9-9457-71dd656403df tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "337d88a2-30b9-4846-929e-042bd7a64a65" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.421s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.317035] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 796.317035] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 796.317035] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Deleting the datastore file [datastore1] 06067572-a9fd-43a4-91cd-383a62f53885 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 796.317472] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4d074daa-45dd-4255-8698-14f01dc50453 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.334860] env[62569]: DEBUG oslo_vmware.api [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for the task: (returnval){ [ 796.334860] env[62569]: value = "task-1250050" [ 796.334860] env[62569]: _type = "Task" [ 796.334860] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.341885] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250044, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.762025} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.346148] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] eb70341b-4282-4eca-b6a2-374db7c521c5/eb70341b-4282-4eca-b6a2-374db7c521c5.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 796.346417] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 796.350096] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fe648899-f9f4-437a-8236-ba6afe6a64ab {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.352286] env[62569]: DEBUG oslo_vmware.api [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250048, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.360829} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.353103] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 796.353330] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 796.353532] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 796.353715] env[62569]: INFO nova.compute.manager [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Took 1.44 seconds to destroy the instance on the hypervisor. [ 796.353970] env[62569]: DEBUG oslo.service.loopingcall [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 796.354553] env[62569]: DEBUG nova.compute.manager [-] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 796.354677] env[62569]: DEBUG nova.network.neutron [-] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 796.362183] env[62569]: DEBUG oslo_vmware.api [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250050, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.362524] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 796.362524] env[62569]: value = "task-1250051" [ 796.362524] env[62569]: _type = "Task" [ 796.362524] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.372732] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250051, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.588786] env[62569]: DEBUG oslo_vmware.api [None req-bec307ea-8c18-4ffc-92e1-ba2a0f707041 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250046, 'name': ReconfigVM_Task, 'duration_secs': 0.851457} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.589087] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-bec307ea-8c18-4ffc-92e1-ba2a0f707041 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Reconfigured VM instance instance-00000034 to attach disk [datastore1] volume-025e7ad1-7cda-4045-99fb-506305a5e0cf/volume-025e7ad1-7cda-4045-99fb-506305a5e0cf.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 796.594387] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4cbf9762-3161-48ec-b853-6a15fcf15ea0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.617031] env[62569]: DEBUG oslo_vmware.api [None req-bec307ea-8c18-4ffc-92e1-ba2a0f707041 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 796.617031] env[62569]: value = "task-1250052" [ 796.617031] env[62569]: _type = "Task" [ 796.617031] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.625301] env[62569]: DEBUG oslo_vmware.api [None req-bec307ea-8c18-4ffc-92e1-ba2a0f707041 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250052, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.765219] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Releasing lock "refresh_cache-94392d48-223d-4205-9627-25488a468769" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.765720] env[62569]: DEBUG nova.compute.manager [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 796.765896] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 796.766940] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600614aa-0a86-4664-af12-968e878e011f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.775714] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 796.778376] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2f12b3c0-8789-448d-aaea-147f98fa02cd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.784500] env[62569]: DEBUG oslo_vmware.api [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 796.784500] env[62569]: value = "task-1250053" [ 796.784500] env[62569]: _type = "Task" [ 796.784500] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.795542] env[62569]: DEBUG oslo_vmware.api [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1250053, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.803213] env[62569]: DEBUG nova.compute.manager [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 796.851533] env[62569]: DEBUG oslo_vmware.api [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Task: {'id': task-1250050, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.348664} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.854181] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 796.854637] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 796.854851] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 796.855063] env[62569]: INFO nova.compute.manager [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Took 1.15 seconds to destroy the instance on the hypervisor. [ 796.855372] env[62569]: DEBUG oslo.service.loopingcall [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 796.856316] env[62569]: DEBUG nova.compute.manager [-] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 796.856416] env[62569]: DEBUG nova.network.neutron [-] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 796.876327] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250051, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.135296} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.876689] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 796.877565] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88bfa17d-c993-4fbe-881d-18c8204fabaa {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.910883] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] eb70341b-4282-4eca-b6a2-374db7c521c5/eb70341b-4282-4eca-b6a2-374db7c521c5.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 796.917229] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9627113-3c40-42d5-8d65-4c08f9f2676d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.938950] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 796.938950] env[62569]: value = "task-1250054" [ 796.938950] env[62569]: _type = "Task" [ 796.938950] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.949040] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250054, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.005092] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64ac06b9-bdaa-43f6-9e69-0e24829831db {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.013368] env[62569]: DEBUG oslo_vmware.rw_handles [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5249d8dd-86c7-d2ae-a6e1-36b2dd94a3bd/disk-0.vmdk. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 797.015060] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f438f38-661b-4a95-8e9d-8ae5e68c80a3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.018498] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1271b61-9d73-4c3b-a4e5-366a03886a1a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.023993] env[62569]: DEBUG oslo_vmware.rw_handles [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5249d8dd-86c7-d2ae-a6e1-36b2dd94a3bd/disk-0.vmdk is in state: ready. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 797.024210] env[62569]: ERROR oslo_vmware.rw_handles [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5249d8dd-86c7-d2ae-a6e1-36b2dd94a3bd/disk-0.vmdk due to incomplete transfer. [ 797.048822] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-cb59aac6-d57e-4c2d-8bb9-ae02424b9c1a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.051156] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed34767c-6306-41d1-af19-518fe286b11e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.060859] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39334e52-5f65-44d0-a9e7-963df6d13067 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.066060] env[62569]: DEBUG oslo_vmware.rw_handles [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5249d8dd-86c7-d2ae-a6e1-36b2dd94a3bd/disk-0.vmdk. {{(pid=62569) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 797.066374] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Uploaded image e370b729-2673-4bfd-b92d-2a4671f01d25 to the Glance image server {{(pid=62569) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 797.068858] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Destroying the VM {{(pid=62569) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 797.069146] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c0a49642-2de0-42f9-9c23-f420c4c11c54 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.081105] env[62569]: DEBUG nova.compute.provider_tree [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 797.083681] env[62569]: DEBUG oslo_vmware.api [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 797.083681] env[62569]: value = "task-1250055" [ 797.083681] env[62569]: _type = "Task" [ 797.083681] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.091953] env[62569]: DEBUG oslo_vmware.api [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250055, 'name': Destroy_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.126734] env[62569]: DEBUG oslo_vmware.api [None req-bec307ea-8c18-4ffc-92e1-ba2a0f707041 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250052, 'name': ReconfigVM_Task, 'duration_secs': 0.178832} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.127177] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-bec307ea-8c18-4ffc-92e1-ba2a0f707041 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269399', 'volume_id': '025e7ad1-7cda-4045-99fb-506305a5e0cf', 'name': 'volume-025e7ad1-7cda-4045-99fb-506305a5e0cf', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a0b406c3-9466-41bd-9de1-e675cab2ceef', 'attached_at': '', 'detached_at': '', 'volume_id': '025e7ad1-7cda-4045-99fb-506305a5e0cf', 'serial': '025e7ad1-7cda-4045-99fb-506305a5e0cf'} {{(pid=62569) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 797.134404] env[62569]: DEBUG nova.network.neutron [-] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.206114] env[62569]: DEBUG nova.compute.manager [req-2078e607-73c7-4367-8ab3-af9a4f07c887 req-6af039a2-e63d-413d-af72-761f18d8759c service nova] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Received event network-vif-deleted-2a1e0442-6ff2-4954-8798-fd35ba7b59d5 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 797.295407] env[62569]: DEBUG oslo_vmware.api [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1250053, 'name': PowerOffVM_Task, 'duration_secs': 0.17123} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.295766] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 797.296036] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 797.296372] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-95766cb4-8224-485b-94e5-6fc38b342363 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.321389] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 797.321730] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 797.321880] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Deleting the datastore file [datastore1] 94392d48-223d-4205-9627-25488a468769 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 797.322182] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4fdb96b2-d312-414c-b913-7a7416eca5cf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.328014] env[62569]: DEBUG oslo_concurrency.lockutils [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.332920] env[62569]: DEBUG oslo_vmware.api [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for the task: (returnval){ [ 797.332920] env[62569]: value = "task-1250057" [ 797.332920] env[62569]: _type = "Task" [ 797.332920] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.341746] env[62569]: DEBUG oslo_vmware.api [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1250057, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.449153] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250054, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.596492] env[62569]: DEBUG oslo_vmware.api [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250055, 'name': Destroy_Task, 'duration_secs': 0.334358} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.596783] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Destroyed the VM [ 797.597190] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Deleting Snapshot of the VM instance {{(pid=62569) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 797.597466] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-00b5b932-15f3-4b47-9f39-795ac5bb4a28 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.603245] env[62569]: DEBUG oslo_vmware.api [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 797.603245] env[62569]: value = "task-1250058" [ 797.603245] env[62569]: _type = "Task" [ 797.603245] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.611596] env[62569]: DEBUG oslo_vmware.api [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250058, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.634911] env[62569]: INFO nova.compute.manager [-] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Took 1.28 seconds to deallocate network for instance. [ 797.637986] env[62569]: DEBUG nova.scheduler.client.report [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Updated inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with generation 82 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 797.637986] env[62569]: DEBUG nova.compute.provider_tree [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Updating resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 generation from 82 to 83 during operation: update_inventory {{(pid=62569) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 797.637986] env[62569]: DEBUG nova.compute.provider_tree [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 797.641030] env[62569]: DEBUG nova.network.neutron [-] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.851023] env[62569]: DEBUG oslo_vmware.api [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Task: {'id': task-1250057, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.319735} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.851023] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 797.851023] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 797.851023] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 797.851023] env[62569]: INFO nova.compute.manager [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] [instance: 94392d48-223d-4205-9627-25488a468769] Took 1.08 seconds to destroy the instance on the hypervisor. [ 797.851362] env[62569]: DEBUG oslo.service.loopingcall [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 797.851362] env[62569]: DEBUG nova.compute.manager [-] [instance: 94392d48-223d-4205-9627-25488a468769] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 797.851362] env[62569]: DEBUG nova.network.neutron [-] [instance: 94392d48-223d-4205-9627-25488a468769] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 797.876383] env[62569]: DEBUG nova.network.neutron [-] [instance: 94392d48-223d-4205-9627-25488a468769] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 797.951166] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250054, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.114686] env[62569]: DEBUG oslo_vmware.api [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250058, 'name': RemoveSnapshot_Task, 'duration_secs': 0.436934} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.114959] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Deleted Snapshot of the VM instance {{(pid=62569) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 798.115221] env[62569]: INFO nova.compute.manager [None req-71fc6fcc-6bf2-4c72-95eb-d3dab1df0431 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Took 15.32 seconds to snapshot the instance on the hypervisor. [ 798.143374] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.541s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.143914] env[62569]: DEBUG nova.compute.manager [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 798.146734] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.375s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.148483] env[62569]: INFO nova.compute.claims [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 798.151508] env[62569]: INFO nova.compute.manager [-] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Took 1.30 seconds to deallocate network for instance. [ 798.152266] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.191030] env[62569]: DEBUG nova.objects.instance [None req-bec307ea-8c18-4ffc-92e1-ba2a0f707041 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lazy-loading 'flavor' on Instance uuid a0b406c3-9466-41bd-9de1-e675cab2ceef {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 798.354227] env[62569]: DEBUG nova.compute.manager [req-fbb0ad8e-4dd7-4744-b821-860dc0cb8b33 req-09ef068f-a4db-4dea-bc89-c5b07747f257 service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Received event network-changed-303f46d4-6e0a-418c-b62b-17323ab645b6 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 798.354474] env[62569]: DEBUG nova.compute.manager [req-fbb0ad8e-4dd7-4744-b821-860dc0cb8b33 req-09ef068f-a4db-4dea-bc89-c5b07747f257 service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Refreshing instance network info cache due to event network-changed-303f46d4-6e0a-418c-b62b-17323ab645b6. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 798.354709] env[62569]: DEBUG oslo_concurrency.lockutils [req-fbb0ad8e-4dd7-4744-b821-860dc0cb8b33 req-09ef068f-a4db-4dea-bc89-c5b07747f257 service nova] Acquiring lock "refresh_cache-337d88a2-30b9-4846-929e-042bd7a64a65" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.354892] env[62569]: DEBUG oslo_concurrency.lockutils [req-fbb0ad8e-4dd7-4744-b821-860dc0cb8b33 req-09ef068f-a4db-4dea-bc89-c5b07747f257 service nova] Acquired lock "refresh_cache-337d88a2-30b9-4846-929e-042bd7a64a65" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.355503] env[62569]: DEBUG nova.network.neutron [req-fbb0ad8e-4dd7-4744-b821-860dc0cb8b33 req-09ef068f-a4db-4dea-bc89-c5b07747f257 service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Refreshing network info cache for port 303f46d4-6e0a-418c-b62b-17323ab645b6 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 798.380509] env[62569]: DEBUG nova.network.neutron [-] [instance: 94392d48-223d-4205-9627-25488a468769] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.450693] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250054, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.654058] env[62569]: DEBUG nova.compute.utils [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 798.655425] env[62569]: DEBUG nova.compute.manager [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 798.655601] env[62569]: DEBUG nova.network.neutron [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 798.660096] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.698994] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bec307ea-8c18-4ffc-92e1-ba2a0f707041 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "a0b406c3-9466-41bd-9de1-e675cab2ceef" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.817s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.863776] env[62569]: DEBUG nova.policy [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de926d5954a148d0be18f305c53e4495', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dbde8d8f33cc4cb5a05c0f94f872d6c2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 798.882826] env[62569]: INFO nova.compute.manager [-] [instance: 94392d48-223d-4205-9627-25488a468769] Took 1.03 seconds to deallocate network for instance. [ 798.925275] env[62569]: DEBUG oslo_concurrency.lockutils [None req-feea2c66-3529-4b3f-811a-bd65dc90078e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "a0b406c3-9466-41bd-9de1-e675cab2ceef" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.925561] env[62569]: DEBUG oslo_concurrency.lockutils [None req-feea2c66-3529-4b3f-811a-bd65dc90078e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "a0b406c3-9466-41bd-9de1-e675cab2ceef" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.925792] env[62569]: DEBUG nova.compute.manager [None req-feea2c66-3529-4b3f-811a-bd65dc90078e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 798.927407] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3468d9e-8770-4ebf-b185-e96affd6a753 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.933803] env[62569]: DEBUG nova.compute.manager [None req-feea2c66-3529-4b3f-811a-bd65dc90078e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62569) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 798.934378] env[62569]: DEBUG nova.objects.instance [None req-feea2c66-3529-4b3f-811a-bd65dc90078e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lazy-loading 'flavor' on Instance uuid a0b406c3-9466-41bd-9de1-e675cab2ceef {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 798.950928] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250054, 'name': ReconfigVM_Task, 'duration_secs': 1.817687} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.951303] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Reconfigured VM instance instance-00000042 to attach disk [datastore1] eb70341b-4282-4eca-b6a2-374db7c521c5/eb70341b-4282-4eca-b6a2-374db7c521c5.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 798.952590] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ad5f7ee0-c54d-4cfe-990d-d4d89cad5060 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.959894] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 798.959894] env[62569]: value = "task-1250059" [ 798.959894] env[62569]: _type = "Task" [ 798.959894] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.970174] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250059, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.153523] env[62569]: DEBUG nova.network.neutron [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Successfully created port: 8c024324-0c17-4626-984c-a4e3309e1b44 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 799.161923] env[62569]: DEBUG nova.compute.manager [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 799.361166] env[62569]: DEBUG nova.network.neutron [req-fbb0ad8e-4dd7-4744-b821-860dc0cb8b33 req-09ef068f-a4db-4dea-bc89-c5b07747f257 service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Updated VIF entry in instance network info cache for port 303f46d4-6e0a-418c-b62b-17323ab645b6. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 799.361421] env[62569]: DEBUG nova.network.neutron [req-fbb0ad8e-4dd7-4744-b821-860dc0cb8b33 req-09ef068f-a4db-4dea-bc89-c5b07747f257 service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Updating instance_info_cache with network_info: [{"id": "303f46d4-6e0a-418c-b62b-17323ab645b6", "address": "fa:16:3e:68:ca:b8", "network": {"id": "497ea2f3-6685-4d07-9952-b5ccf3c0b2f8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-561818925-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c69ac971dbd419f870a9fa68850cb1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7654928b-7afe-42e3-a18d-68ecc775cefe", "external-id": "cl2-zone-807", "segmentation_id": 807, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap303f46d4-6e", "ovs_interfaceid": "303f46d4-6e0a-418c-b62b-17323ab645b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.388785] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.436250] env[62569]: DEBUG nova.compute.manager [req-5fe663ac-c851-4e89-b24c-31fd49d4baec req-933e16f8-55b7-4cc0-a144-55112716be60 service nova] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Received event network-vif-deleted-1ae266f0-188a-45de-9a09-d9e21b48bff3 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 799.478381] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250059, 'name': Rename_Task, 'duration_secs': 0.278559} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.479210] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 799.480190] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7a957a15-7a56-44db-8c6d-601020e88794 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.486921] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 799.486921] env[62569]: value = "task-1250060" [ 799.486921] env[62569]: _type = "Task" [ 799.486921] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.497961] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250060, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.528254] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f139e45-0b07-4dc8-83a8-ae918ad44032 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.537779] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b8c940-6dce-4a03-b2a1-9a934bfdf927 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.567607] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58f8033e-d0a8-4d48-8b3e-ee905fb38042 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.574869] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-250bc2f7-d8ab-4602-9b15-b79b3f9ad22d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.588946] env[62569]: DEBUG nova.compute.provider_tree [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 799.865486] env[62569]: DEBUG oslo_concurrency.lockutils [req-fbb0ad8e-4dd7-4744-b821-860dc0cb8b33 req-09ef068f-a4db-4dea-bc89-c5b07747f257 service nova] Releasing lock "refresh_cache-337d88a2-30b9-4846-929e-042bd7a64a65" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.943424] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-feea2c66-3529-4b3f-811a-bd65dc90078e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 799.943769] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a95dd2bb-5a47-4c1c-9234-77fcae956d3d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.951289] env[62569]: DEBUG oslo_vmware.api [None req-feea2c66-3529-4b3f-811a-bd65dc90078e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 799.951289] env[62569]: value = "task-1250061" [ 799.951289] env[62569]: _type = "Task" [ 799.951289] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.961363] env[62569]: DEBUG oslo_vmware.api [None req-feea2c66-3529-4b3f-811a-bd65dc90078e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250061, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.996161] env[62569]: DEBUG oslo_vmware.api [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250060, 'name': PowerOnVM_Task, 'duration_secs': 0.467012} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.996471] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 799.996708] env[62569]: INFO nova.compute.manager [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Took 9.31 seconds to spawn the instance on the hypervisor. [ 799.996910] env[62569]: DEBUG nova.compute.manager [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 799.997705] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccc877b3-ed7a-4ee9-be84-4af73b01a0bf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.126332] env[62569]: DEBUG nova.scheduler.client.report [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Updated inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with generation 83 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 800.126612] env[62569]: DEBUG nova.compute.provider_tree [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Updating resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 generation from 83 to 84 during operation: update_inventory {{(pid=62569) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 800.126797] env[62569]: DEBUG nova.compute.provider_tree [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 800.180233] env[62569]: DEBUG nova.compute.manager [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 800.194327] env[62569]: DEBUG oslo_concurrency.lockutils [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "22094c32-5f50-4f86-a77b-cd4adcf8998a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.194586] env[62569]: DEBUG oslo_concurrency.lockutils [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "22094c32-5f50-4f86-a77b-cd4adcf8998a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.194798] env[62569]: DEBUG oslo_concurrency.lockutils [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "22094c32-5f50-4f86-a77b-cd4adcf8998a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.194985] env[62569]: DEBUG oslo_concurrency.lockutils [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "22094c32-5f50-4f86-a77b-cd4adcf8998a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.195175] env[62569]: DEBUG oslo_concurrency.lockutils [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "22094c32-5f50-4f86-a77b-cd4adcf8998a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.200734] env[62569]: INFO nova.compute.manager [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Terminating instance [ 800.282826] env[62569]: DEBUG nova.virt.hardware [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 800.283091] env[62569]: DEBUG nova.virt.hardware [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 800.283253] env[62569]: DEBUG nova.virt.hardware [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 800.283435] env[62569]: DEBUG nova.virt.hardware [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 800.283594] env[62569]: DEBUG nova.virt.hardware [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 800.283760] env[62569]: DEBUG nova.virt.hardware [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 800.283969] env[62569]: DEBUG nova.virt.hardware [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 800.284228] env[62569]: DEBUG nova.virt.hardware [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 800.284328] env[62569]: DEBUG nova.virt.hardware [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 800.284495] env[62569]: DEBUG nova.virt.hardware [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 800.284743] env[62569]: DEBUG nova.virt.hardware [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 800.286589] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a05d7d44-e796-4859-b4b0-c1f630685ac1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.294839] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f34f82bb-5700-45e1-ba77-f7719319a312 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.462082] env[62569]: DEBUG oslo_vmware.api [None req-feea2c66-3529-4b3f-811a-bd65dc90078e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250061, 'name': PowerOffVM_Task, 'duration_secs': 0.423444} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.462082] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-feea2c66-3529-4b3f-811a-bd65dc90078e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 800.462082] env[62569]: DEBUG nova.compute.manager [None req-feea2c66-3529-4b3f-811a-bd65dc90078e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 800.462638] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a2542dd-da54-466a-a4fa-dd96e594359a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.472589] env[62569]: DEBUG nova.compute.manager [req-6b0496dc-0f0e-4856-85ff-4ee8ab30d349 req-0b724922-9624-4390-8824-5112f1601391 service nova] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Received event network-changed-06154dac-fb91-4e11-a503-34ae0c293331 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 800.472654] env[62569]: DEBUG nova.compute.manager [req-6b0496dc-0f0e-4856-85ff-4ee8ab30d349 req-0b724922-9624-4390-8824-5112f1601391 service nova] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Refreshing instance network info cache due to event network-changed-06154dac-fb91-4e11-a503-34ae0c293331. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 800.472891] env[62569]: DEBUG oslo_concurrency.lockutils [req-6b0496dc-0f0e-4856-85ff-4ee8ab30d349 req-0b724922-9624-4390-8824-5112f1601391 service nova] Acquiring lock "refresh_cache-b3ccc28b-68d0-461d-a67d-b5ad179a80f9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.473056] env[62569]: DEBUG oslo_concurrency.lockutils [req-6b0496dc-0f0e-4856-85ff-4ee8ab30d349 req-0b724922-9624-4390-8824-5112f1601391 service nova] Acquired lock "refresh_cache-b3ccc28b-68d0-461d-a67d-b5ad179a80f9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.473257] env[62569]: DEBUG nova.network.neutron [req-6b0496dc-0f0e-4856-85ff-4ee8ab30d349 req-0b724922-9624-4390-8824-5112f1601391 service nova] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Refreshing network info cache for port 06154dac-fb91-4e11-a503-34ae0c293331 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 800.515933] env[62569]: INFO nova.compute.manager [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Took 36.12 seconds to build instance. [ 800.600050] env[62569]: DEBUG oslo_concurrency.lockutils [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Acquiring lock "b3ccc28b-68d0-461d-a67d-b5ad179a80f9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.601071] env[62569]: DEBUG oslo_concurrency.lockutils [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Lock "b3ccc28b-68d0-461d-a67d-b5ad179a80f9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.603401] env[62569]: DEBUG oslo_concurrency.lockutils [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Acquiring lock "b3ccc28b-68d0-461d-a67d-b5ad179a80f9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.603401] env[62569]: DEBUG oslo_concurrency.lockutils [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Lock "b3ccc28b-68d0-461d-a67d-b5ad179a80f9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.603401] env[62569]: DEBUG oslo_concurrency.lockutils [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Lock "b3ccc28b-68d0-461d-a67d-b5ad179a80f9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.604354] env[62569]: INFO nova.compute.manager [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Terminating instance [ 800.632714] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.485s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.632714] env[62569]: DEBUG nova.compute.manager [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 800.635504] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.456s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.639816] env[62569]: INFO nova.compute.claims [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 800.706658] env[62569]: DEBUG nova.compute.manager [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 800.706901] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 800.707806] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-100f3188-1405-4b9e-9e95-657a230f8725 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.715945] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 800.716173] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3c5cd054-0341-4d14-b04f-f6a4f2ac5d5d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.728232] env[62569]: DEBUG nova.compute.manager [req-79ee5c8d-c1e6-4561-b841-dfec24285d8e req-9c1f515f-c69a-4b77-a980-f71a8a1f18e1 service nova] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Received event network-vif-plugged-8c024324-0c17-4626-984c-a4e3309e1b44 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 800.728232] env[62569]: DEBUG oslo_concurrency.lockutils [req-79ee5c8d-c1e6-4561-b841-dfec24285d8e req-9c1f515f-c69a-4b77-a980-f71a8a1f18e1 service nova] Acquiring lock "cefbca2e-609d-4954-bec6-52ffe095446f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.728232] env[62569]: DEBUG oslo_concurrency.lockutils [req-79ee5c8d-c1e6-4561-b841-dfec24285d8e req-9c1f515f-c69a-4b77-a980-f71a8a1f18e1 service nova] Lock "cefbca2e-609d-4954-bec6-52ffe095446f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.728232] env[62569]: DEBUG oslo_concurrency.lockutils [req-79ee5c8d-c1e6-4561-b841-dfec24285d8e req-9c1f515f-c69a-4b77-a980-f71a8a1f18e1 service nova] Lock "cefbca2e-609d-4954-bec6-52ffe095446f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.728232] env[62569]: DEBUG nova.compute.manager [req-79ee5c8d-c1e6-4561-b841-dfec24285d8e req-9c1f515f-c69a-4b77-a980-f71a8a1f18e1 service nova] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] No waiting events found dispatching network-vif-plugged-8c024324-0c17-4626-984c-a4e3309e1b44 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 800.728657] env[62569]: WARNING nova.compute.manager [req-79ee5c8d-c1e6-4561-b841-dfec24285d8e req-9c1f515f-c69a-4b77-a980-f71a8a1f18e1 service nova] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Received unexpected event network-vif-plugged-8c024324-0c17-4626-984c-a4e3309e1b44 for instance with vm_state building and task_state spawning. [ 800.869925] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "05db90a6-3faf-4878-b782-ed17df47ed5f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.871035] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "05db90a6-3faf-4878-b782-ed17df47ed5f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.979969] env[62569]: DEBUG oslo_concurrency.lockutils [None req-feea2c66-3529-4b3f-811a-bd65dc90078e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "a0b406c3-9466-41bd-9de1-e675cab2ceef" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.054s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.019639] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8fee41ef-df1e-4d44-b7bd-64fb88aafe08 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "eb70341b-4282-4eca-b6a2-374db7c521c5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.664s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.110744] env[62569]: DEBUG nova.compute.manager [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 801.110744] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 801.110744] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9e3fb79-1606-419a-9402-905896c868a3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.118401] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 801.118852] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-badf4ac5-6638-419c-a080-c1074effec16 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.127842] env[62569]: DEBUG oslo_vmware.api [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Waiting for the task: (returnval){ [ 801.127842] env[62569]: value = "task-1250063" [ 801.127842] env[62569]: _type = "Task" [ 801.127842] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.136928] env[62569]: DEBUG oslo_vmware.api [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Task: {'id': task-1250063, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.141620] env[62569]: DEBUG nova.compute.utils [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 801.144885] env[62569]: DEBUG nova.compute.manager [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 801.145240] env[62569]: DEBUG nova.network.neutron [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 801.205448] env[62569]: DEBUG nova.network.neutron [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Successfully updated port: 8c024324-0c17-4626-984c-a4e3309e1b44 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 801.220015] env[62569]: DEBUG nova.policy [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ab6d20862c54432cbafdda33ccc974ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '530ca7e7924743ab91a362a064a3111b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 801.431775] env[62569]: DEBUG nova.network.neutron [req-6b0496dc-0f0e-4856-85ff-4ee8ab30d349 req-0b724922-9624-4390-8824-5112f1601391 service nova] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Updated VIF entry in instance network info cache for port 06154dac-fb91-4e11-a503-34ae0c293331. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 801.432242] env[62569]: DEBUG nova.network.neutron [req-6b0496dc-0f0e-4856-85ff-4ee8ab30d349 req-0b724922-9624-4390-8824-5112f1601391 service nova] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Updating instance_info_cache with network_info: [{"id": "06154dac-fb91-4e11-a503-34ae0c293331", "address": "fa:16:3e:ab:d8:01", "network": {"id": "9c4e9b79-be5c-4b09-b717-0f2b7423bc05", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1323416416-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7633be3f1b38470dae7048754b408f0c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d69a4b11-8d65-435f-94a5-28f74a39a718", "external-id": "cl2-zone-59", "segmentation_id": 59, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06154dac-fb", "ovs_interfaceid": "06154dac-fb91-4e11-a503-34ae0c293331", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.523584] env[62569]: DEBUG nova.compute.manager [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 801.527675] env[62569]: DEBUG nova.network.neutron [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Successfully created port: 9d855102-33ab-4574-b791-dfa6fb5b0f09 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 801.637354] env[62569]: DEBUG oslo_vmware.api [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Task: {'id': task-1250063, 'name': PowerOffVM_Task, 'duration_secs': 0.323739} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.637721] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 801.637807] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 801.638062] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d3f612e0-e63f-4526-8e14-58633137af79 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.650526] env[62569]: DEBUG nova.compute.manager [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 801.662605] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 801.662836] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 801.663062] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Deleting the datastore file [datastore1] 22094c32-5f50-4f86-a77b-cd4adcf8998a {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 801.663462] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-699b87a2-d086-457f-b2f6-69bf9f734631 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.672088] env[62569]: DEBUG oslo_vmware.api [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 801.672088] env[62569]: value = "task-1250065" [ 801.672088] env[62569]: _type = "Task" [ 801.672088] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.683811] env[62569]: DEBUG oslo_vmware.api [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250065, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.707420] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Acquiring lock "refresh_cache-cefbca2e-609d-4954-bec6-52ffe095446f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.707568] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Acquired lock "refresh_cache-cefbca2e-609d-4954-bec6-52ffe095446f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.707740] env[62569]: DEBUG nova.network.neutron [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 801.727392] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 801.727665] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 801.727876] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Deleting the datastore file [datastore2] b3ccc28b-68d0-461d-a67d-b5ad179a80f9 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 801.730660] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5123b246-07e9-45f2-85fb-eefa78ade439 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.739446] env[62569]: DEBUG oslo_vmware.api [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Waiting for the task: (returnval){ [ 801.739446] env[62569]: value = "task-1250066" [ 801.739446] env[62569]: _type = "Task" [ 801.739446] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.750948] env[62569]: DEBUG oslo_vmware.api [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Task: {'id': task-1250066, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.804064] env[62569]: DEBUG nova.compute.manager [req-a6414703-27c2-4db4-adc0-ed35b828bfb9 req-d3af0ed1-fee2-4ad8-8150-4124e8dc7d95 service nova] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Received event network-changed-8c024324-0c17-4626-984c-a4e3309e1b44 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 801.804340] env[62569]: DEBUG nova.compute.manager [req-a6414703-27c2-4db4-adc0-ed35b828bfb9 req-d3af0ed1-fee2-4ad8-8150-4124e8dc7d95 service nova] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Refreshing instance network info cache due to event network-changed-8c024324-0c17-4626-984c-a4e3309e1b44. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 801.804529] env[62569]: DEBUG oslo_concurrency.lockutils [req-a6414703-27c2-4db4-adc0-ed35b828bfb9 req-d3af0ed1-fee2-4ad8-8150-4124e8dc7d95 service nova] Acquiring lock "refresh_cache-cefbca2e-609d-4954-bec6-52ffe095446f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.934819] env[62569]: DEBUG oslo_concurrency.lockutils [req-6b0496dc-0f0e-4856-85ff-4ee8ab30d349 req-0b724922-9624-4390-8824-5112f1601391 service nova] Releasing lock "refresh_cache-b3ccc28b-68d0-461d-a67d-b5ad179a80f9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.008123] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1b4dab9-b20f-42cf-b794-93880a9f5726 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.016610] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f517f684-866a-4c44-af5e-d1ab1240df4f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.051868] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8f4b747-ec05-48ec-8d19-b84e76dd96ee {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.062461] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-214b9078-7b2f-456c-9baa-d01cf9647f0f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.073621] env[62569]: DEBUG nova.compute.provider_tree [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.080971] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.200198] env[62569]: DEBUG oslo_vmware.api [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250065, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159131} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.200700] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 802.201337] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 802.201871] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 802.202216] env[62569]: INFO nova.compute.manager [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Took 1.50 seconds to destroy the instance on the hypervisor. [ 802.202623] env[62569]: DEBUG oslo.service.loopingcall [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 802.203318] env[62569]: DEBUG nova.compute.manager [-] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 802.203508] env[62569]: DEBUG nova.network.neutron [-] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 802.253316] env[62569]: DEBUG oslo_vmware.api [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Task: {'id': task-1250066, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131981} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.254461] env[62569]: DEBUG nova.network.neutron [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 802.257056] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 802.257394] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 802.257723] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 802.258046] env[62569]: INFO nova.compute.manager [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Took 1.15 seconds to destroy the instance on the hypervisor. [ 802.258447] env[62569]: DEBUG oslo.service.loopingcall [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 802.258774] env[62569]: DEBUG nova.compute.manager [-] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 802.258963] env[62569]: DEBUG nova.network.neutron [-] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 802.393865] env[62569]: DEBUG nova.network.neutron [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Updating instance_info_cache with network_info: [{"id": "8c024324-0c17-4626-984c-a4e3309e1b44", "address": "fa:16:3e:7f:54:f9", "network": {"id": "99adcab7-798d-4135-9fc0-681b4cd0f71f", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1843124593-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbde8d8f33cc4cb5a05c0f94f872d6c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "05b1253d-2b87-4158-9ff1-dafcf829f11f", "external-id": "nsx-vlan-transportzone-55", "segmentation_id": 55, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c024324-0c", "ovs_interfaceid": "8c024324-0c17-4626-984c-a4e3309e1b44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.448584] env[62569]: DEBUG nova.objects.instance [None req-190c3fe5-149c-498f-bd0a-25deafd06a0e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lazy-loading 'flavor' on Instance uuid a0b406c3-9466-41bd-9de1-e675cab2ceef {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 802.580709] env[62569]: DEBUG nova.scheduler.client.report [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 802.682156] env[62569]: DEBUG nova.compute.manager [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 802.685448] env[62569]: INFO nova.compute.manager [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Rebuilding instance [ 802.717211] env[62569]: DEBUG nova.virt.hardware [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 802.717497] env[62569]: DEBUG nova.virt.hardware [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.717664] env[62569]: DEBUG nova.virt.hardware [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 802.717855] env[62569]: DEBUG nova.virt.hardware [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.718149] env[62569]: DEBUG nova.virt.hardware [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 802.722508] env[62569]: DEBUG nova.virt.hardware [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 802.722846] env[62569]: DEBUG nova.virt.hardware [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 802.725411] env[62569]: DEBUG nova.virt.hardware [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 802.725411] env[62569]: DEBUG nova.virt.hardware [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 802.725411] env[62569]: DEBUG nova.virt.hardware [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 802.725411] env[62569]: DEBUG nova.virt.hardware [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 802.726741] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddbfdb9d-e031-4b64-b961-d7d63602e451 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.735157] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28b1cdad-0591-465b-b72d-199551255524 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.762510] env[62569]: DEBUG nova.compute.manager [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 802.763496] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aac694a1-5795-4640-9198-208b21c7f30f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.896483] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Releasing lock "refresh_cache-cefbca2e-609d-4954-bec6-52ffe095446f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.898415] env[62569]: DEBUG nova.compute.manager [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Instance network_info: |[{"id": "8c024324-0c17-4626-984c-a4e3309e1b44", "address": "fa:16:3e:7f:54:f9", "network": {"id": "99adcab7-798d-4135-9fc0-681b4cd0f71f", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1843124593-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbde8d8f33cc4cb5a05c0f94f872d6c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "05b1253d-2b87-4158-9ff1-dafcf829f11f", "external-id": "nsx-vlan-transportzone-55", "segmentation_id": 55, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c024324-0c", "ovs_interfaceid": "8c024324-0c17-4626-984c-a4e3309e1b44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 802.898415] env[62569]: DEBUG oslo_concurrency.lockutils [req-a6414703-27c2-4db4-adc0-ed35b828bfb9 req-d3af0ed1-fee2-4ad8-8150-4124e8dc7d95 service nova] Acquired lock "refresh_cache-cefbca2e-609d-4954-bec6-52ffe095446f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.898699] env[62569]: DEBUG nova.network.neutron [req-a6414703-27c2-4db4-adc0-ed35b828bfb9 req-d3af0ed1-fee2-4ad8-8150-4124e8dc7d95 service nova] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Refreshing network info cache for port 8c024324-0c17-4626-984c-a4e3309e1b44 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 802.899420] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7f:54:f9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '05b1253d-2b87-4158-9ff1-dafcf829f11f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8c024324-0c17-4626-984c-a4e3309e1b44', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 802.907943] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Creating folder: Project (dbde8d8f33cc4cb5a05c0f94f872d6c2). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 802.912751] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-58873f9f-9fbf-4870-b701-5313e1a2dc76 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.924381] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Created folder: Project (dbde8d8f33cc4cb5a05c0f94f872d6c2) in parent group-v269330. [ 802.924688] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Creating folder: Instances. Parent ref: group-v269403. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 802.925091] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ca2d6ae5-3617-4e29-8a64-4b74dde27d4a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.933897] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Created folder: Instances in parent group-v269403. [ 802.934243] env[62569]: DEBUG oslo.service.loopingcall [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 802.934498] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 802.934768] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8a78fb37-3a24-4598-bbd7-30547e314f98 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.952903] env[62569]: DEBUG oslo_concurrency.lockutils [None req-190c3fe5-149c-498f-bd0a-25deafd06a0e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "refresh_cache-a0b406c3-9466-41bd-9de1-e675cab2ceef" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.953187] env[62569]: DEBUG oslo_concurrency.lockutils [None req-190c3fe5-149c-498f-bd0a-25deafd06a0e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquired lock "refresh_cache-a0b406c3-9466-41bd-9de1-e675cab2ceef" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.953412] env[62569]: DEBUG nova.network.neutron [None req-190c3fe5-149c-498f-bd0a-25deafd06a0e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 802.953634] env[62569]: DEBUG nova.objects.instance [None req-190c3fe5-149c-498f-bd0a-25deafd06a0e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lazy-loading 'info_cache' on Instance uuid a0b406c3-9466-41bd-9de1-e675cab2ceef {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 802.956188] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 802.956188] env[62569]: value = "task-1250069" [ 802.956188] env[62569]: _type = "Task" [ 802.956188] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.967334] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250069, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.085423] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.450s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.085991] env[62569]: DEBUG nova.compute.manager [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 803.093475] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.427s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.095124] env[62569]: INFO nova.compute.claims [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 803.265233] env[62569]: DEBUG nova.network.neutron [req-a6414703-27c2-4db4-adc0-ed35b828bfb9 req-d3af0ed1-fee2-4ad8-8150-4124e8dc7d95 service nova] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Updated VIF entry in instance network info cache for port 8c024324-0c17-4626-984c-a4e3309e1b44. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 803.265957] env[62569]: DEBUG nova.network.neutron [req-a6414703-27c2-4db4-adc0-ed35b828bfb9 req-d3af0ed1-fee2-4ad8-8150-4124e8dc7d95 service nova] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Updating instance_info_cache with network_info: [{"id": "8c024324-0c17-4626-984c-a4e3309e1b44", "address": "fa:16:3e:7f:54:f9", "network": {"id": "99adcab7-798d-4135-9fc0-681b4cd0f71f", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1843124593-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dbde8d8f33cc4cb5a05c0f94f872d6c2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "05b1253d-2b87-4158-9ff1-dafcf829f11f", "external-id": "nsx-vlan-transportzone-55", "segmentation_id": 55, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c024324-0c", "ovs_interfaceid": "8c024324-0c17-4626-984c-a4e3309e1b44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.293170] env[62569]: DEBUG nova.network.neutron [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Successfully updated port: 9d855102-33ab-4574-b791-dfa6fb5b0f09 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 803.459127] env[62569]: DEBUG nova.objects.base [None req-190c3fe5-149c-498f-bd0a-25deafd06a0e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62569) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 803.474832] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250069, 'name': CreateVM_Task, 'duration_secs': 0.376731} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.476679] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 803.477528] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.477699] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.478332] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 803.479046] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f091b737-1f7b-4cce-939b-41781c2b1a7f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.485075] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Waiting for the task: (returnval){ [ 803.485075] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f0cfb9-df8e-14ea-ccdd-1f94b13aedef" [ 803.485075] env[62569]: _type = "Task" [ 803.485075] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.492250] env[62569]: DEBUG nova.compute.manager [req-2cbc0a2f-7cb4-4469-a59d-7d07d094f9b9 req-92e19e86-51d7-429c-8843-4978f3c93fd9 service nova] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Received event network-vif-deleted-06154dac-fb91-4e11-a503-34ae0c293331 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 803.492458] env[62569]: INFO nova.compute.manager [req-2cbc0a2f-7cb4-4469-a59d-7d07d094f9b9 req-92e19e86-51d7-429c-8843-4978f3c93fd9 service nova] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Neutron deleted interface 06154dac-fb91-4e11-a503-34ae0c293331; detaching it from the instance and deleting it from the info cache [ 803.492635] env[62569]: DEBUG nova.network.neutron [req-2cbc0a2f-7cb4-4469-a59d-7d07d094f9b9 req-92e19e86-51d7-429c-8843-4978f3c93fd9 service nova] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.500837] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f0cfb9-df8e-14ea-ccdd-1f94b13aedef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.605684] env[62569]: DEBUG nova.compute.utils [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 803.607319] env[62569]: DEBUG nova.compute.manager [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 803.607533] env[62569]: DEBUG nova.network.neutron [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 803.620271] env[62569]: DEBUG nova.network.neutron [-] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.649773] env[62569]: DEBUG nova.policy [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4613deb515db4042b02af229bb05e9ca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ed0f5952cbd45eaa31512d6df93dc9e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 803.677522] env[62569]: DEBUG nova.network.neutron [-] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.773692] env[62569]: DEBUG oslo_concurrency.lockutils [req-a6414703-27c2-4db4-adc0-ed35b828bfb9 req-d3af0ed1-fee2-4ad8-8150-4124e8dc7d95 service nova] Releasing lock "refresh_cache-cefbca2e-609d-4954-bec6-52ffe095446f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.778558] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 803.778853] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-041b5ba0-e208-4efd-b1d2-bfa2164838f1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.786488] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 803.786488] env[62569]: value = "task-1250070" [ 803.786488] env[62569]: _type = "Task" [ 803.786488] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.794836] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250070, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.795566] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.795748] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.796009] env[62569]: DEBUG nova.network.neutron [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 803.940805] env[62569]: DEBUG nova.compute.manager [req-f309921d-dedc-4239-aa91-7370da390b32 req-104cbedf-2730-43c9-89d0-d4af83cbe8c0 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Received event network-vif-plugged-9d855102-33ab-4574-b791-dfa6fb5b0f09 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 803.941228] env[62569]: DEBUG oslo_concurrency.lockutils [req-f309921d-dedc-4239-aa91-7370da390b32 req-104cbedf-2730-43c9-89d0-d4af83cbe8c0 service nova] Acquiring lock "807d3025-d6a7-4778-a829-a61e2c7495c9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.941365] env[62569]: DEBUG oslo_concurrency.lockutils [req-f309921d-dedc-4239-aa91-7370da390b32 req-104cbedf-2730-43c9-89d0-d4af83cbe8c0 service nova] Lock "807d3025-d6a7-4778-a829-a61e2c7495c9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.941548] env[62569]: DEBUG oslo_concurrency.lockutils [req-f309921d-dedc-4239-aa91-7370da390b32 req-104cbedf-2730-43c9-89d0-d4af83cbe8c0 service nova] Lock "807d3025-d6a7-4778-a829-a61e2c7495c9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.941940] env[62569]: DEBUG nova.compute.manager [req-f309921d-dedc-4239-aa91-7370da390b32 req-104cbedf-2730-43c9-89d0-d4af83cbe8c0 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] No waiting events found dispatching network-vif-plugged-9d855102-33ab-4574-b791-dfa6fb5b0f09 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 803.941940] env[62569]: WARNING nova.compute.manager [req-f309921d-dedc-4239-aa91-7370da390b32 req-104cbedf-2730-43c9-89d0-d4af83cbe8c0 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Received unexpected event network-vif-plugged-9d855102-33ab-4574-b791-dfa6fb5b0f09 for instance with vm_state building and task_state spawning. [ 803.942281] env[62569]: DEBUG nova.compute.manager [req-f309921d-dedc-4239-aa91-7370da390b32 req-104cbedf-2730-43c9-89d0-d4af83cbe8c0 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Received event network-changed-9d855102-33ab-4574-b791-dfa6fb5b0f09 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 803.942412] env[62569]: DEBUG nova.compute.manager [req-f309921d-dedc-4239-aa91-7370da390b32 req-104cbedf-2730-43c9-89d0-d4af83cbe8c0 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Refreshing instance network info cache due to event network-changed-9d855102-33ab-4574-b791-dfa6fb5b0f09. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 803.942592] env[62569]: DEBUG oslo_concurrency.lockutils [req-f309921d-dedc-4239-aa91-7370da390b32 req-104cbedf-2730-43c9-89d0-d4af83cbe8c0 service nova] Acquiring lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.996766] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f0cfb9-df8e-14ea-ccdd-1f94b13aedef, 'name': SearchDatastore_Task, 'duration_secs': 0.039182} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.997351] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.997469] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 804.001352] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.001546] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.001760] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 804.002343] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c583a500-92f4-4ac7-9811-5a9e66c3bfda {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.004999] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a5e5880b-8c9b-4ee1-9c12-b7b8ffd80b21 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.024417] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d61cc9f-02a9-4004-9556-19dc0d9fbcc1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.034774] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 804.034966] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 804.035814] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-856c04f1-1571-46ea-8c80-3ac4cd1478b4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.042567] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Waiting for the task: (returnval){ [ 804.042567] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52580b5b-f54b-74aa-27e0-4595aff1e32b" [ 804.042567] env[62569]: _type = "Task" [ 804.042567] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.049971] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52580b5b-f54b-74aa-27e0-4595aff1e32b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.056617] env[62569]: DEBUG nova.compute.manager [req-2cbc0a2f-7cb4-4469-a59d-7d07d094f9b9 req-92e19e86-51d7-429c-8843-4978f3c93fd9 service nova] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Detach interface failed, port_id=06154dac-fb91-4e11-a503-34ae0c293331, reason: Instance b3ccc28b-68d0-461d-a67d-b5ad179a80f9 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 804.061912] env[62569]: DEBUG nova.network.neutron [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Successfully created port: a466b80c-87f5-4bfd-ad2a-fc3610a634ed {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 804.111461] env[62569]: DEBUG nova.compute.manager [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 804.123855] env[62569]: INFO nova.compute.manager [-] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Took 1.86 seconds to deallocate network for instance. [ 804.184809] env[62569]: INFO nova.compute.manager [-] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Took 1.98 seconds to deallocate network for instance. [ 804.300553] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250070, 'name': PowerOffVM_Task, 'duration_secs': 0.180903} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.302770] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 804.303197] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 804.304655] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-745520f8-8913-4d5e-9414-228ce950fc6b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.312315] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 804.312949] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f2db6abc-fc65-484f-a846-6dc5f9d1d97a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.352847] env[62569]: DEBUG nova.network.neutron [None req-190c3fe5-149c-498f-bd0a-25deafd06a0e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Updating instance_info_cache with network_info: [{"id": "3da67b46-e89d-4a96-8dd9-752956b9b270", "address": "fa:16:3e:f6:2c:c2", "network": {"id": "ba6b71b1-bd7b-493e-8682-68d842fa935f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1892479706-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e9ea55b00b574a7db0530efcb498ca2b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3da67b46-e8", "ovs_interfaceid": "3da67b46-e89d-4a96-8dd9-752956b9b270", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.363181] env[62569]: DEBUG nova.network.neutron [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 804.420691] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 804.421232] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 804.421232] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Deleting the datastore file [datastore1] eb70341b-4282-4eca-b6a2-374db7c521c5 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 804.421363] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-45955505-6e77-48d0-b4f5-785cdd3b039d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.430236] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 804.430236] env[62569]: value = "task-1250072" [ 804.430236] env[62569]: _type = "Task" [ 804.430236] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.441623] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250072, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.446059] env[62569]: DEBUG nova.network.neutron [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Successfully created port: ac9a07c1-60a1-4fd4-83c7-54ffc30240d8 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 804.493623] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17530203-92f0-4012-bea0-12e9ec7684ac {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.502026] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daf90d01-ca56-4e4a-8f6b-7e310f334aa0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.544955] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bea1635c-e027-4e42-8784-361a3cf3e716 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.557237] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52580b5b-f54b-74aa-27e0-4595aff1e32b, 'name': SearchDatastore_Task, 'duration_secs': 0.009764} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.560229] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14e76033-98ab-49d6-9f10-4b1be0145400 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.566245] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9f2f976-ce53-499d-8b64-cc83ac8386e6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.571315] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Waiting for the task: (returnval){ [ 804.571315] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52342f82-a569-2836-3a12-254396d1b80c" [ 804.571315] env[62569]: _type = "Task" [ 804.571315] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.579574] env[62569]: DEBUG nova.compute.provider_tree [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.589283] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52342f82-a569-2836-3a12-254396d1b80c, 'name': SearchDatastore_Task, 'duration_secs': 0.009091} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.589283] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.589534] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] cefbca2e-609d-4954-bec6-52ffe095446f/cefbca2e-609d-4954-bec6-52ffe095446f.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 804.590362] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8a0ea009-2297-4d3f-b565-4515f26c64f6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.597041] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Waiting for the task: (returnval){ [ 804.597041] env[62569]: value = "task-1250073" [ 804.597041] env[62569]: _type = "Task" [ 804.597041] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.605235] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Task: {'id': task-1250073, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.634025] env[62569]: DEBUG oslo_concurrency.lockutils [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.670153] env[62569]: DEBUG nova.network.neutron [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Updating instance_info_cache with network_info: [{"id": "9d855102-33ab-4574-b791-dfa6fb5b0f09", "address": "fa:16:3e:b3:4b:f5", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d855102-33", "ovs_interfaceid": "9d855102-33ab-4574-b791-dfa6fb5b0f09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.693668] env[62569]: DEBUG oslo_concurrency.lockutils [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.854665] env[62569]: DEBUG oslo_concurrency.lockutils [None req-190c3fe5-149c-498f-bd0a-25deafd06a0e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Releasing lock "refresh_cache-a0b406c3-9466-41bd-9de1-e675cab2ceef" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.943995] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250072, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.129189} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.943995] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 804.944144] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 804.944272] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 805.085449] env[62569]: DEBUG nova.scheduler.client.report [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 805.107021] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Task: {'id': task-1250073, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.44565} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.109668] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] cefbca2e-609d-4954-bec6-52ffe095446f/cefbca2e-609d-4954-bec6-52ffe095446f.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 805.109668] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 805.109668] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f82a7ab6-08eb-4e83-80ba-b775ab871c51 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.114532] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Waiting for the task: (returnval){ [ 805.114532] env[62569]: value = "task-1250074" [ 805.114532] env[62569]: _type = "Task" [ 805.114532] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.121716] env[62569]: DEBUG nova.compute.manager [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 805.127275] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Task: {'id': task-1250074, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.147304] env[62569]: DEBUG nova.virt.hardware [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 805.147807] env[62569]: DEBUG nova.virt.hardware [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 805.148240] env[62569]: DEBUG nova.virt.hardware [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 805.149029] env[62569]: DEBUG nova.virt.hardware [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 805.149029] env[62569]: DEBUG nova.virt.hardware [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 805.149029] env[62569]: DEBUG nova.virt.hardware [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 805.149186] env[62569]: DEBUG nova.virt.hardware [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 805.149302] env[62569]: DEBUG nova.virt.hardware [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 805.149482] env[62569]: DEBUG nova.virt.hardware [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 805.149649] env[62569]: DEBUG nova.virt.hardware [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 805.149840] env[62569]: DEBUG nova.virt.hardware [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 805.150885] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e5ea5c7-7584-4ecd-9bd8-b9c4ec7bb892 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.160293] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3370346e-7372-426a-9763-e729c1e95562 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.173774] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.174110] env[62569]: DEBUG nova.compute.manager [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Instance network_info: |[{"id": "9d855102-33ab-4574-b791-dfa6fb5b0f09", "address": "fa:16:3e:b3:4b:f5", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d855102-33", "ovs_interfaceid": "9d855102-33ab-4574-b791-dfa6fb5b0f09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 805.174667] env[62569]: DEBUG oslo_concurrency.lockutils [req-f309921d-dedc-4239-aa91-7370da390b32 req-104cbedf-2730-43c9-89d0-d4af83cbe8c0 service nova] Acquired lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.174910] env[62569]: DEBUG nova.network.neutron [req-f309921d-dedc-4239-aa91-7370da390b32 req-104cbedf-2730-43c9-89d0-d4af83cbe8c0 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Refreshing network info cache for port 9d855102-33ab-4574-b791-dfa6fb5b0f09 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 805.176125] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b3:4b:f5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '099fe970-c61f-4480-bed4-ae4f485fd82a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9d855102-33ab-4574-b791-dfa6fb5b0f09', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 805.183655] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Creating folder: Project (530ca7e7924743ab91a362a064a3111b). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 805.184699] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bcf746c7-d307-42d2-bba7-1dee29ab803f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.194040] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Created folder: Project (530ca7e7924743ab91a362a064a3111b) in parent group-v269330. [ 805.194040] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Creating folder: Instances. Parent ref: group-v269406. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 805.194213] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ef9588c6-ca80-4d2d-ba9f-4ca1f725ad91 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.202640] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Created folder: Instances in parent group-v269406. [ 805.202889] env[62569]: DEBUG oslo.service.loopingcall [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 805.203099] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 805.203310] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-720bf2e7-fad2-4594-8592-424067aa6145 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.221462] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 805.221462] env[62569]: value = "task-1250077" [ 805.221462] env[62569]: _type = "Task" [ 805.221462] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.228466] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250077, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.537865] env[62569]: DEBUG nova.compute.manager [req-e45a683b-4d6b-4381-8c10-3d59b51c91d4 req-58901105-d4b9-4ffd-8a15-43e826415593 service nova] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Received event network-vif-deleted-e1705802-dcd4-48a2-bad5-ffbd36806564 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 805.591092] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.498s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.591510] env[62569]: DEBUG nova.compute.manager [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 805.596870] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.394s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.598891] env[62569]: INFO nova.compute.claims [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 805.627823] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Task: {'id': task-1250074, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063306} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.628327] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 805.631955] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d518c354-2471-4252-9e61-9e455f829b06 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.657843] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] cefbca2e-609d-4954-bec6-52ffe095446f/cefbca2e-609d-4954-bec6-52ffe095446f.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 805.658815] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-74ab10c7-df99-4b91-90f9-8a56ee59fb80 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.680090] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Waiting for the task: (returnval){ [ 805.680090] env[62569]: value = "task-1250078" [ 805.680090] env[62569]: _type = "Task" [ 805.680090] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.690939] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Task: {'id': task-1250078, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.735079] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250077, 'name': CreateVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.865375] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-190c3fe5-149c-498f-bd0a-25deafd06a0e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 805.865375] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-98a73f36-bfb8-400a-821c-dd8bc98cfaae {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.872020] env[62569]: DEBUG oslo_vmware.api [None req-190c3fe5-149c-498f-bd0a-25deafd06a0e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 805.872020] env[62569]: value = "task-1250079" [ 805.872020] env[62569]: _type = "Task" [ 805.872020] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.879465] env[62569]: DEBUG oslo_vmware.api [None req-190c3fe5-149c-498f-bd0a-25deafd06a0e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250079, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.945436] env[62569]: DEBUG nova.network.neutron [req-f309921d-dedc-4239-aa91-7370da390b32 req-104cbedf-2730-43c9-89d0-d4af83cbe8c0 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Updated VIF entry in instance network info cache for port 9d855102-33ab-4574-b791-dfa6fb5b0f09. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 805.945921] env[62569]: DEBUG nova.network.neutron [req-f309921d-dedc-4239-aa91-7370da390b32 req-104cbedf-2730-43c9-89d0-d4af83cbe8c0 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Updating instance_info_cache with network_info: [{"id": "9d855102-33ab-4574-b791-dfa6fb5b0f09", "address": "fa:16:3e:b3:4b:f5", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d855102-33", "ovs_interfaceid": "9d855102-33ab-4574-b791-dfa6fb5b0f09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.987018] env[62569]: DEBUG nova.virt.hardware [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 805.987018] env[62569]: DEBUG nova.virt.hardware [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 805.987018] env[62569]: DEBUG nova.virt.hardware [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 805.987239] env[62569]: DEBUG nova.virt.hardware [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 805.987239] env[62569]: DEBUG nova.virt.hardware [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 805.987239] env[62569]: DEBUG nova.virt.hardware [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 805.987239] env[62569]: DEBUG nova.virt.hardware [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 805.987239] env[62569]: DEBUG nova.virt.hardware [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 805.987399] env[62569]: DEBUG nova.virt.hardware [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 805.987399] env[62569]: DEBUG nova.virt.hardware [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 805.987399] env[62569]: DEBUG nova.virt.hardware [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 805.987399] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5857e86c-6747-40b5-91ba-cd3eaabcb1f1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.994565] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da7922f6-e183-4fd2-b33a-82afda3408a9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.000817] env[62569]: DEBUG nova.compute.manager [req-a8665cb3-95ee-41e8-aa94-d347ac1d1b49 req-ccd488a2-9407-495c-ac9d-016826e88d67 service nova] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Received event network-vif-plugged-a466b80c-87f5-4bfd-ad2a-fc3610a634ed {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 806.001041] env[62569]: DEBUG oslo_concurrency.lockutils [req-a8665cb3-95ee-41e8-aa94-d347ac1d1b49 req-ccd488a2-9407-495c-ac9d-016826e88d67 service nova] Acquiring lock "390bcf25-689d-46ad-bffb-3670c3729397-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.001266] env[62569]: DEBUG oslo_concurrency.lockutils [req-a8665cb3-95ee-41e8-aa94-d347ac1d1b49 req-ccd488a2-9407-495c-ac9d-016826e88d67 service nova] Lock "390bcf25-689d-46ad-bffb-3670c3729397-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.001436] env[62569]: DEBUG oslo_concurrency.lockutils [req-a8665cb3-95ee-41e8-aa94-d347ac1d1b49 req-ccd488a2-9407-495c-ac9d-016826e88d67 service nova] Lock "390bcf25-689d-46ad-bffb-3670c3729397-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.001618] env[62569]: DEBUG nova.compute.manager [req-a8665cb3-95ee-41e8-aa94-d347ac1d1b49 req-ccd488a2-9407-495c-ac9d-016826e88d67 service nova] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] No waiting events found dispatching network-vif-plugged-a466b80c-87f5-4bfd-ad2a-fc3610a634ed {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 806.001789] env[62569]: WARNING nova.compute.manager [req-a8665cb3-95ee-41e8-aa94-d347ac1d1b49 req-ccd488a2-9407-495c-ac9d-016826e88d67 service nova] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Received unexpected event network-vif-plugged-a466b80c-87f5-4bfd-ad2a-fc3610a634ed for instance with vm_state building and task_state spawning. [ 806.012995] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:15:fa:ec', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6685c85e-be1e-4b7b-a6cc-3e50e59b6567', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'abd9f1dd-a26a-4f38-bee3-041710cbd705', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 806.022480] env[62569]: DEBUG oslo.service.loopingcall [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 806.023230] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 806.023436] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-03c5b290-f163-4bc5-bfbb-e1c3ba917670 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.045563] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 806.045563] env[62569]: value = "task-1250080" [ 806.045563] env[62569]: _type = "Task" [ 806.045563] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.053944] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250080, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.097637] env[62569]: DEBUG nova.compute.utils [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 806.099154] env[62569]: DEBUG nova.compute.manager [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 806.099330] env[62569]: DEBUG nova.network.neutron [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 806.172317] env[62569]: DEBUG nova.policy [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9beb6ad713b3458e9bcbbf28fca0d6ed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b3de498c4aa74ba698deedcacb9b44f4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 806.194198] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Task: {'id': task-1250078, 'name': ReconfigVM_Task, 'duration_secs': 0.308638} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.194582] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Reconfigured VM instance instance-00000043 to attach disk [datastore1] cefbca2e-609d-4954-bec6-52ffe095446f/cefbca2e-609d-4954-bec6-52ffe095446f.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 806.195304] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-475fdf48-6c36-4181-9c0d-4a8c4ccce550 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.201758] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Waiting for the task: (returnval){ [ 806.201758] env[62569]: value = "task-1250081" [ 806.201758] env[62569]: _type = "Task" [ 806.201758] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.210084] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Task: {'id': task-1250081, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.232232] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250077, 'name': CreateVM_Task, 'duration_secs': 0.532318} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.232401] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 806.233199] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.233365] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.233691] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 806.233952] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0527bf61-02d2-40eb-9f3c-10a6499124ad {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.238765] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 806.238765] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b7d560-43b1-62ad-7482-7d330cde60ee" [ 806.238765] env[62569]: _type = "Task" [ 806.238765] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.248113] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b7d560-43b1-62ad-7482-7d330cde60ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.383014] env[62569]: DEBUG oslo_vmware.api [None req-190c3fe5-149c-498f-bd0a-25deafd06a0e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250079, 'name': PowerOnVM_Task, 'duration_secs': 0.416181} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.383014] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-190c3fe5-149c-498f-bd0a-25deafd06a0e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 806.383014] env[62569]: DEBUG nova.compute.manager [None req-190c3fe5-149c-498f-bd0a-25deafd06a0e tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 806.383014] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-957b4715-8ed5-4fed-a70f-e1d3e9e7107a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.449925] env[62569]: DEBUG oslo_concurrency.lockutils [req-f309921d-dedc-4239-aa91-7370da390b32 req-104cbedf-2730-43c9-89d0-d4af83cbe8c0 service nova] Releasing lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.557386] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250080, 'name': CreateVM_Task, 'duration_secs': 0.408726} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.557578] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 806.558291] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.558416] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.558746] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 806.559036] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e474e71f-715d-44c2-9398-6abd416ec84d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.566055] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 806.566055] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52e5c835-642e-61be-daca-0233b0858668" [ 806.566055] env[62569]: _type = "Task" [ 806.566055] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.571256] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52e5c835-642e-61be-daca-0233b0858668, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.603250] env[62569]: DEBUG nova.compute.manager [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 806.633701] env[62569]: DEBUG nova.network.neutron [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Successfully updated port: a466b80c-87f5-4bfd-ad2a-fc3610a634ed {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 806.713088] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Task: {'id': task-1250081, 'name': Rename_Task, 'duration_secs': 0.240652} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.713088] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 806.713910] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a1e397e2-93ed-4d20-8c19-bd72f45f76ee {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.723241] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Waiting for the task: (returnval){ [ 806.723241] env[62569]: value = "task-1250082" [ 806.723241] env[62569]: _type = "Task" [ 806.723241] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.738156] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Task: {'id': task-1250082, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.753507] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b7d560-43b1-62ad-7482-7d330cde60ee, 'name': SearchDatastore_Task, 'duration_secs': 0.023088} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.753836] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.754227] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 806.754479] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.754656] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.755018] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 806.757736] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c6a6d42e-9344-4dcf-b514-9988a767c423 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.767441] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 806.767643] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 806.768717] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b33ea768-b9ae-4387-913b-33baf1fead64 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.777532] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 806.777532] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]526ea239-f059-6d94-e133-0b84a015b817" [ 806.777532] env[62569]: _type = "Task" [ 806.777532] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.787223] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]526ea239-f059-6d94-e133-0b84a015b817, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.942550] env[62569]: DEBUG nova.compute.manager [req-28207d1f-648f-4620-ae3e-23a1c4387fb4 req-c1b137aa-1b9e-462f-baec-2c40aa58793a service nova] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Received event network-changed-a466b80c-87f5-4bfd-ad2a-fc3610a634ed {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 806.942761] env[62569]: DEBUG nova.compute.manager [req-28207d1f-648f-4620-ae3e-23a1c4387fb4 req-c1b137aa-1b9e-462f-baec-2c40aa58793a service nova] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Refreshing instance network info cache due to event network-changed-a466b80c-87f5-4bfd-ad2a-fc3610a634ed. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 806.942978] env[62569]: DEBUG oslo_concurrency.lockutils [req-28207d1f-648f-4620-ae3e-23a1c4387fb4 req-c1b137aa-1b9e-462f-baec-2c40aa58793a service nova] Acquiring lock "refresh_cache-390bcf25-689d-46ad-bffb-3670c3729397" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.944948] env[62569]: DEBUG oslo_concurrency.lockutils [req-28207d1f-648f-4620-ae3e-23a1c4387fb4 req-c1b137aa-1b9e-462f-baec-2c40aa58793a service nova] Acquired lock "refresh_cache-390bcf25-689d-46ad-bffb-3670c3729397" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.945204] env[62569]: DEBUG nova.network.neutron [req-28207d1f-648f-4620-ae3e-23a1c4387fb4 req-c1b137aa-1b9e-462f-baec-2c40aa58793a service nova] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Refreshing network info cache for port a466b80c-87f5-4bfd-ad2a-fc3610a634ed {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 806.952358] env[62569]: DEBUG nova.network.neutron [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Successfully created port: da0de7a0-922e-46b2-842b-a222e31de3b2 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 806.990301] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78a86528-83fb-47d7-8538-dcf20b4a138e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.998232] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4fe0e47-5f7f-4c15-a199-a8cc3220ff07 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.034794] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de37b7b-a437-4f15-8a72-50ce4bb25e3f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.044017] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4183ff5-bcd9-41bb-b0f2-1fe9679f6b8e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.059917] env[62569]: DEBUG nova.compute.provider_tree [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 807.073711] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52e5c835-642e-61be-daca-0233b0858668, 'name': SearchDatastore_Task, 'duration_secs': 0.039047} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.074035] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.074332] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 807.074515] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.074669] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.074883] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 807.075138] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1666984f-8fc9-4891-abab-452681cd13d8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.084315] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 807.084518] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 807.085310] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fb2a4c4-9b28-48f3-8745-65bef9be4900 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.090882] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 807.090882] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b41938-ab67-6f8a-5bc0-9947eb85e527" [ 807.090882] env[62569]: _type = "Task" [ 807.090882] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.099541] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b41938-ab67-6f8a-5bc0-9947eb85e527, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.108524] env[62569]: INFO nova.virt.block_device [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Booting with volume 3fb6b4c3-cf28-44c4-9606-7430ec94b7f6 at /dev/sda [ 807.154150] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ba8cbdc0-308c-439d-9e56-cf31f4cfbae6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.165019] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e17b2e72-96da-4dc2-ace8-290dd23635cc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.192375] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b4200e3b-48af-4c94-8e62-82b49bc68d78 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.200937] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ed3d25a-1d6a-44e5-a760-425a2e01f785 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.233439] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ec62bf9-5102-4649-8afc-5878f292e5ab {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.244329] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Task: {'id': task-1250082, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.247114] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cdba03a-99a5-459f-80e9-046018941680 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.260951] env[62569]: DEBUG nova.virt.block_device [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Updating existing volume attachment record: df0f16e4-d160-4dca-94f7-563d020d5b2b {{(pid=62569) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 807.290469] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]526ea239-f059-6d94-e133-0b84a015b817, 'name': SearchDatastore_Task, 'duration_secs': 0.011612} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.291324] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b6c5de7-9490-4d42-b2d9-d48705d3c1cc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.297418] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 807.297418] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52730732-8a28-22f0-c8c8-ef6948005f6c" [ 807.297418] env[62569]: _type = "Task" [ 807.297418] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.306095] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52730732-8a28-22f0-c8c8-ef6948005f6c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.507517] env[62569]: DEBUG nova.network.neutron [req-28207d1f-648f-4620-ae3e-23a1c4387fb4 req-c1b137aa-1b9e-462f-baec-2c40aa58793a service nova] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 807.564539] env[62569]: DEBUG nova.scheduler.client.report [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 807.602984] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b41938-ab67-6f8a-5bc0-9947eb85e527, 'name': SearchDatastore_Task, 'duration_secs': 0.009093} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.603787] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9c6e1ae-def3-4923-af61-a16bc1e41486 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.608855] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 807.608855] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5225c4f2-e42b-bccb-3c38-e550ab44b2a7" [ 807.608855] env[62569]: _type = "Task" [ 807.608855] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.619026] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5225c4f2-e42b-bccb-3c38-e550ab44b2a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.642234] env[62569]: DEBUG nova.network.neutron [req-28207d1f-648f-4620-ae3e-23a1c4387fb4 req-c1b137aa-1b9e-462f-baec-2c40aa58793a service nova] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.742806] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Task: {'id': task-1250082, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.809499] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52730732-8a28-22f0-c8c8-ef6948005f6c, 'name': SearchDatastore_Task, 'duration_secs': 0.018469} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.811101] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.811400] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 807d3025-d6a7-4778-a829-a61e2c7495c9/807d3025-d6a7-4778-a829-a61e2c7495c9.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 807.811855] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f1ec56cc-ff2e-4eec-88f0-90286a6b6cc2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.819211] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 807.819211] env[62569]: value = "task-1250083" [ 807.819211] env[62569]: _type = "Task" [ 807.819211] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.830633] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250083, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.070717] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.475s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.071380] env[62569]: DEBUG nova.compute.manager [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 808.075790] env[62569]: DEBUG oslo_concurrency.lockutils [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.510s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.076486] env[62569]: DEBUG nova.objects.instance [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Lazy-loading 'resources' on Instance uuid 4feb7b78-9f7c-4e64-b0a7-870ed73adf97 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 808.123218] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5225c4f2-e42b-bccb-3c38-e550ab44b2a7, 'name': SearchDatastore_Task, 'duration_secs': 0.01179} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.123218] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.123652] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] eb70341b-4282-4eca-b6a2-374db7c521c5/eb70341b-4282-4eca-b6a2-374db7c521c5.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 808.123766] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0b33f7de-3bc8-4599-810e-a6e375cd6b2e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.132448] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 808.132448] env[62569]: value = "task-1250084" [ 808.132448] env[62569]: _type = "Task" [ 808.132448] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.142451] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250084, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.145194] env[62569]: DEBUG oslo_concurrency.lockutils [req-28207d1f-648f-4620-ae3e-23a1c4387fb4 req-c1b137aa-1b9e-462f-baec-2c40aa58793a service nova] Releasing lock "refresh_cache-390bcf25-689d-46ad-bffb-3670c3729397" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.241869] env[62569]: DEBUG oslo_vmware.api [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Task: {'id': task-1250082, 'name': PowerOnVM_Task, 'duration_secs': 1.369862} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.242240] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 808.242490] env[62569]: INFO nova.compute.manager [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Took 8.06 seconds to spawn the instance on the hypervisor. [ 808.242703] env[62569]: DEBUG nova.compute.manager [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 808.243595] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0326e25-13b4-4b5b-9529-4ec07fe4031d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.330232] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250083, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476096} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.330596] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 807d3025-d6a7-4778-a829-a61e2c7495c9/807d3025-d6a7-4778-a829-a61e2c7495c9.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 808.330845] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 808.331148] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-25dfe271-7e38-46ff-ac6b-70b4ff7fa377 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.337709] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 808.337709] env[62569]: value = "task-1250085" [ 808.337709] env[62569]: _type = "Task" [ 808.337709] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.345488] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250085, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.511344] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "c8edf395-ea21-4ba8-991e-0f43fe4ee830" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.511586] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "c8edf395-ea21-4ba8-991e-0f43fe4ee830" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.579671] env[62569]: DEBUG nova.compute.utils [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 808.585060] env[62569]: DEBUG nova.compute.manager [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 808.585304] env[62569]: DEBUG nova.network.neutron [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 808.654596] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250084, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.656222] env[62569]: DEBUG nova.policy [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b9e0da546a03426782fcacd06a7afa8a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '00bdb4c3f88f4d61834f7e961629ef3a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 808.772146] env[62569]: INFO nova.compute.manager [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Took 32.66 seconds to build instance. [ 808.773615] env[62569]: DEBUG nova.compute.manager [req-22394be4-a690-40bb-8d10-7d08e02df2f0 req-3e2acfbf-bf22-41ca-84eb-c9969f1e41ae service nova] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Received event network-vif-plugged-da0de7a0-922e-46b2-842b-a222e31de3b2 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 808.774494] env[62569]: DEBUG oslo_concurrency.lockutils [req-22394be4-a690-40bb-8d10-7d08e02df2f0 req-3e2acfbf-bf22-41ca-84eb-c9969f1e41ae service nova] Acquiring lock "fdeac8b8-a34d-4b4c-8f72-281f84e22ebe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.774773] env[62569]: DEBUG oslo_concurrency.lockutils [req-22394be4-a690-40bb-8d10-7d08e02df2f0 req-3e2acfbf-bf22-41ca-84eb-c9969f1e41ae service nova] Lock "fdeac8b8-a34d-4b4c-8f72-281f84e22ebe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.774954] env[62569]: DEBUG oslo_concurrency.lockutils [req-22394be4-a690-40bb-8d10-7d08e02df2f0 req-3e2acfbf-bf22-41ca-84eb-c9969f1e41ae service nova] Lock "fdeac8b8-a34d-4b4c-8f72-281f84e22ebe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.775145] env[62569]: DEBUG nova.compute.manager [req-22394be4-a690-40bb-8d10-7d08e02df2f0 req-3e2acfbf-bf22-41ca-84eb-c9969f1e41ae service nova] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] No waiting events found dispatching network-vif-plugged-da0de7a0-922e-46b2-842b-a222e31de3b2 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 808.775312] env[62569]: WARNING nova.compute.manager [req-22394be4-a690-40bb-8d10-7d08e02df2f0 req-3e2acfbf-bf22-41ca-84eb-c9969f1e41ae service nova] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Received unexpected event network-vif-plugged-da0de7a0-922e-46b2-842b-a222e31de3b2 for instance with vm_state building and task_state block_device_mapping. [ 808.848982] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250085, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.125999} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.849285] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 808.850206] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90c5b438-f954-460b-a8f8-1fadab62ab27 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.882959] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 807d3025-d6a7-4778-a829-a61e2c7495c9/807d3025-d6a7-4778-a829-a61e2c7495c9.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 808.885629] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c987acce-0358-4884-888b-40dac3abb003 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.919994] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 808.919994] env[62569]: value = "task-1250086" [ 808.919994] env[62569]: _type = "Task" [ 808.919994] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.929791] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250086, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.954247] env[62569]: DEBUG nova.network.neutron [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Successfully updated port: da0de7a0-922e-46b2-842b-a222e31de3b2 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 809.068225] env[62569]: DEBUG nova.compute.manager [req-4b263374-b74f-47ba-8922-80e909cbbba6 req-d302f1b0-0e71-46be-a0e2-c6d6d0d42c3c service nova] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Received event network-vif-plugged-ac9a07c1-60a1-4fd4-83c7-54ffc30240d8 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 809.068646] env[62569]: DEBUG oslo_concurrency.lockutils [req-4b263374-b74f-47ba-8922-80e909cbbba6 req-d302f1b0-0e71-46be-a0e2-c6d6d0d42c3c service nova] Acquiring lock "390bcf25-689d-46ad-bffb-3670c3729397-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.068975] env[62569]: DEBUG oslo_concurrency.lockutils [req-4b263374-b74f-47ba-8922-80e909cbbba6 req-d302f1b0-0e71-46be-a0e2-c6d6d0d42c3c service nova] Lock "390bcf25-689d-46ad-bffb-3670c3729397-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.069286] env[62569]: DEBUG oslo_concurrency.lockutils [req-4b263374-b74f-47ba-8922-80e909cbbba6 req-d302f1b0-0e71-46be-a0e2-c6d6d0d42c3c service nova] Lock "390bcf25-689d-46ad-bffb-3670c3729397-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.069588] env[62569]: DEBUG nova.compute.manager [req-4b263374-b74f-47ba-8922-80e909cbbba6 req-d302f1b0-0e71-46be-a0e2-c6d6d0d42c3c service nova] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] No waiting events found dispatching network-vif-plugged-ac9a07c1-60a1-4fd4-83c7-54ffc30240d8 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 809.069906] env[62569]: WARNING nova.compute.manager [req-4b263374-b74f-47ba-8922-80e909cbbba6 req-d302f1b0-0e71-46be-a0e2-c6d6d0d42c3c service nova] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Received unexpected event network-vif-plugged-ac9a07c1-60a1-4fd4-83c7-54ffc30240d8 for instance with vm_state building and task_state spawning. [ 809.085819] env[62569]: DEBUG nova.compute.manager [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 809.092127] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d5222d-f425-4bce-8299-1c7ef048c40c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.102136] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9078373-d7a4-4c82-804a-35ea2a98e8f9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.140118] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd6e1be-a49a-45c8-9162-570c964ef421 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.148823] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250084, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.61035} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.151016] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] eb70341b-4282-4eca-b6a2-374db7c521c5/eb70341b-4282-4eca-b6a2-374db7c521c5.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 809.151227] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 809.151586] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-25c6e053-857f-47bd-ad94-9e24a85eace3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.154606] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65ce76ef-96a0-4b41-8fe6-e4f1e4939a5d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.172248] env[62569]: DEBUG nova.compute.provider_tree [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 809.175107] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 809.175107] env[62569]: value = "task-1250087" [ 809.175107] env[62569]: _type = "Task" [ 809.175107] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.183879] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250087, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.209736] env[62569]: DEBUG nova.network.neutron [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Successfully created port: 90b3aaef-2e72-486f-be8d-a4202a6cb0c6 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 809.236320] env[62569]: DEBUG nova.network.neutron [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Successfully updated port: ac9a07c1-60a1-4fd4-83c7-54ffc30240d8 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 809.279612] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0c377d79-5015-4fd6-b4ca-6dc91c2a891e tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Lock "cefbca2e-609d-4954-bec6-52ffe095446f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.112s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.386723] env[62569]: DEBUG nova.compute.manager [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 809.388160] env[62569]: DEBUG nova.virt.hardware [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 809.388460] env[62569]: DEBUG nova.virt.hardware [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 809.388641] env[62569]: DEBUG nova.virt.hardware [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 809.388850] env[62569]: DEBUG nova.virt.hardware [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 809.389016] env[62569]: DEBUG nova.virt.hardware [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 809.389403] env[62569]: DEBUG nova.virt.hardware [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 809.389664] env[62569]: DEBUG nova.virt.hardware [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 809.389849] env[62569]: DEBUG nova.virt.hardware [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 809.390217] env[62569]: DEBUG nova.virt.hardware [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 809.390423] env[62569]: DEBUG nova.virt.hardware [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 809.390637] env[62569]: DEBUG nova.virt.hardware [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 809.391768] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d700ff2c-d2a8-40be-aa55-82945f72a7ff {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.404234] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21ae5e7a-719b-47ea-a7ef-55bfbc20ed00 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.429454] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250086, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.456363] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Acquiring lock "refresh_cache-fdeac8b8-a34d-4b4c-8f72-281f84e22ebe" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.456438] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Acquired lock "refresh_cache-fdeac8b8-a34d-4b4c-8f72-281f84e22ebe" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.456630] env[62569]: DEBUG nova.network.neutron [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 809.678450] env[62569]: DEBUG nova.scheduler.client.report [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 809.694061] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250087, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06573} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.694061] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 809.694689] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d0e081f-677d-42c5-84ef-aeb666b6429c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.728974] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] eb70341b-4282-4eca-b6a2-374db7c521c5/eb70341b-4282-4eca-b6a2-374db7c521c5.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 809.729820] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3ffc3b0-4a1f-43ad-9f28-d8f19aad848a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.748193] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Acquiring lock "refresh_cache-390bcf25-689d-46ad-bffb-3670c3729397" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.748330] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Acquired lock "refresh_cache-390bcf25-689d-46ad-bffb-3670c3729397" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.748488] env[62569]: DEBUG nova.network.neutron [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 809.758788] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 809.758788] env[62569]: value = "task-1250088" [ 809.758788] env[62569]: _type = "Task" [ 809.758788] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.768866] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250088, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.786065] env[62569]: DEBUG nova.compute.manager [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 809.931818] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250086, 'name': ReconfigVM_Task, 'duration_secs': 0.875508} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.932144] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 807d3025-d6a7-4778-a829-a61e2c7495c9/807d3025-d6a7-4778-a829-a61e2c7495c9.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 809.932837] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-118aedf8-c7f1-4881-ad43-cf3c7407297b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.939328] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 809.939328] env[62569]: value = "task-1250089" [ 809.939328] env[62569]: _type = "Task" [ 809.939328] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.948777] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250089, 'name': Rename_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.995704] env[62569]: DEBUG nova.network.neutron [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 810.099703] env[62569]: DEBUG nova.compute.manager [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 810.126751] env[62569]: DEBUG nova.virt.hardware [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 810.126993] env[62569]: DEBUG nova.virt.hardware [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 810.127898] env[62569]: DEBUG nova.virt.hardware [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 810.128341] env[62569]: DEBUG nova.virt.hardware [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 810.128571] env[62569]: DEBUG nova.virt.hardware [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 810.128757] env[62569]: DEBUG nova.virt.hardware [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 810.129019] env[62569]: DEBUG nova.virt.hardware [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 810.129266] env[62569]: DEBUG nova.virt.hardware [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 810.130124] env[62569]: DEBUG nova.virt.hardware [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 810.130124] env[62569]: DEBUG nova.virt.hardware [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 810.130124] env[62569]: DEBUG nova.virt.hardware [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 810.131218] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57229fb3-727a-443f-8fc0-f9c198c60ea8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.140588] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1488461e-79cc-4f19-8d07-3a3f682b027e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.157711] env[62569]: DEBUG nova.network.neutron [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Updating instance_info_cache with network_info: [{"id": "da0de7a0-922e-46b2-842b-a222e31de3b2", "address": "fa:16:3e:eb:02:d2", "network": {"id": "ea8253a1-5878-4093-9acf-1426d966475b", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-616956732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3de498c4aa74ba698deedcacb9b44f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d33839ae-40ca-471b-92e3-eb282b920682", "external-id": "nsx-vlan-transportzone-416", "segmentation_id": 416, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda0de7a0-92", "ovs_interfaceid": "da0de7a0-922e-46b2-842b-a222e31de3b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.191019] env[62569]: DEBUG oslo_concurrency.lockutils [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.113s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.191760] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.369s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.193972] env[62569]: INFO nova.compute.claims [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 810.217031] env[62569]: INFO nova.scheduler.client.report [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Deleted allocations for instance 4feb7b78-9f7c-4e64-b0a7-870ed73adf97 [ 810.269161] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250088, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.288830] env[62569]: DEBUG nova.network.neutron [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 810.303053] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.452038] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250089, 'name': Rename_Task, 'duration_secs': 0.243963} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.452038] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 810.452038] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e66220f7-85a9-4d7e-9e9b-13c92f337c41 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.457679] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 810.457679] env[62569]: value = "task-1250090" [ 810.457679] env[62569]: _type = "Task" [ 810.457679] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.465052] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250090, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.610336] env[62569]: DEBUG nova.network.neutron [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Updating instance_info_cache with network_info: [{"id": "a466b80c-87f5-4bfd-ad2a-fc3610a634ed", "address": "fa:16:3e:38:93:3d", "network": {"id": "fa1d9d59-c210-4299-8da1-8441e1c2595e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-49047811", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.113", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ed0f5952cbd45eaa31512d6df93dc9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa466b80c-87", "ovs_interfaceid": "a466b80c-87f5-4bfd-ad2a-fc3610a634ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "ac9a07c1-60a1-4fd4-83c7-54ffc30240d8", "address": "fa:16:3e:d8:ba:94", "network": {"id": "d2c1889e-76c5-46d6-8d7c-918bf384a898", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1640224982", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.148", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "2ed0f5952cbd45eaa31512d6df93dc9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6e940e5-e083-4238-973e-f1b4e2a3a5c7", "external-id": "nsx-vlan-transportzone-64", "segmentation_id": 64, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac9a07c1-60", "ovs_interfaceid": "ac9a07c1-60a1-4fd4-83c7-54ffc30240d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.660617] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Releasing lock "refresh_cache-fdeac8b8-a34d-4b4c-8f72-281f84e22ebe" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.661020] env[62569]: DEBUG nova.compute.manager [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Instance network_info: |[{"id": "da0de7a0-922e-46b2-842b-a222e31de3b2", "address": "fa:16:3e:eb:02:d2", "network": {"id": "ea8253a1-5878-4093-9acf-1426d966475b", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-616956732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3de498c4aa74ba698deedcacb9b44f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d33839ae-40ca-471b-92e3-eb282b920682", "external-id": "nsx-vlan-transportzone-416", "segmentation_id": 416, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda0de7a0-92", "ovs_interfaceid": "da0de7a0-922e-46b2-842b-a222e31de3b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 810.661459] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:eb:02:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd33839ae-40ca-471b-92e3-eb282b920682', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'da0de7a0-922e-46b2-842b-a222e31de3b2', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 810.668876] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Creating folder: Project (b3de498c4aa74ba698deedcacb9b44f4). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 810.669430] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d9a8f5ce-5260-42f4-a511-09a4c8746f10 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.683042] env[62569]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 810.683210] env[62569]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62569) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 810.683543] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Folder already exists: Project (b3de498c4aa74ba698deedcacb9b44f4). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 810.683748] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Creating folder: Instances. Parent ref: group-v269347. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 810.684006] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e07da00b-9b3a-4180-9eb2-917a111d0187 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.693763] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Created folder: Instances in parent group-v269347. [ 810.694009] env[62569]: DEBUG oslo.service.loopingcall [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 810.694218] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 810.694425] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b7da46cf-3a7f-4103-81ff-6a79526f6292 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.715469] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 810.715469] env[62569]: value = "task-1250093" [ 810.715469] env[62569]: _type = "Task" [ 810.715469] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.725683] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250093, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.726012] env[62569]: DEBUG oslo_concurrency.lockutils [None req-101cf3fe-0605-47d4-84b1-33b39f829d13 tempest-ServerMetadataNegativeTestJSON-506584640 tempest-ServerMetadataNegativeTestJSON-506584640-project-member] Lock "4feb7b78-9f7c-4e64-b0a7-870ed73adf97" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.934s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.772016] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250088, 'name': ReconfigVM_Task, 'duration_secs': 0.580994} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.772754] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Reconfigured VM instance instance-00000042 to attach disk [datastore1] eb70341b-4282-4eca-b6a2-374db7c521c5/eb70341b-4282-4eca-b6a2-374db7c521c5.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 810.773447] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d2420137-33a0-437f-acff-1196f6e88d6e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.780330] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 810.780330] env[62569]: value = "task-1250094" [ 810.780330] env[62569]: _type = "Task" [ 810.780330] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.791754] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250094, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.792563] env[62569]: DEBUG nova.network.neutron [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Successfully updated port: 90b3aaef-2e72-486f-be8d-a4202a6cb0c6 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 810.807994] env[62569]: DEBUG nova.compute.manager [req-f198bf83-4ae2-4c1d-93e0-3ee8e9f00236 req-38f8cf75-16ab-479d-910d-25ae285b8d47 service nova] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Received event network-changed-da0de7a0-922e-46b2-842b-a222e31de3b2 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 810.808142] env[62569]: DEBUG nova.compute.manager [req-f198bf83-4ae2-4c1d-93e0-3ee8e9f00236 req-38f8cf75-16ab-479d-910d-25ae285b8d47 service nova] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Refreshing instance network info cache due to event network-changed-da0de7a0-922e-46b2-842b-a222e31de3b2. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 810.808358] env[62569]: DEBUG oslo_concurrency.lockutils [req-f198bf83-4ae2-4c1d-93e0-3ee8e9f00236 req-38f8cf75-16ab-479d-910d-25ae285b8d47 service nova] Acquiring lock "refresh_cache-fdeac8b8-a34d-4b4c-8f72-281f84e22ebe" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.808604] env[62569]: DEBUG oslo_concurrency.lockutils [req-f198bf83-4ae2-4c1d-93e0-3ee8e9f00236 req-38f8cf75-16ab-479d-910d-25ae285b8d47 service nova] Acquired lock "refresh_cache-fdeac8b8-a34d-4b4c-8f72-281f84e22ebe" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.808828] env[62569]: DEBUG nova.network.neutron [req-f198bf83-4ae2-4c1d-93e0-3ee8e9f00236 req-38f8cf75-16ab-479d-910d-25ae285b8d47 service nova] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Refreshing network info cache for port da0de7a0-922e-46b2-842b-a222e31de3b2 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 810.969360] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250090, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.101919] env[62569]: DEBUG nova.compute.manager [req-9813ccbf-358c-4a80-8eb0-6a2b780f6ef0 req-369592b3-87ef-49ed-8211-5daa3b38905b service nova] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Received event network-changed-ac9a07c1-60a1-4fd4-83c7-54ffc30240d8 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 811.104812] env[62569]: DEBUG nova.compute.manager [req-9813ccbf-358c-4a80-8eb0-6a2b780f6ef0 req-369592b3-87ef-49ed-8211-5daa3b38905b service nova] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Refreshing instance network info cache due to event network-changed-ac9a07c1-60a1-4fd4-83c7-54ffc30240d8. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 811.105173] env[62569]: DEBUG oslo_concurrency.lockutils [req-9813ccbf-358c-4a80-8eb0-6a2b780f6ef0 req-369592b3-87ef-49ed-8211-5daa3b38905b service nova] Acquiring lock "refresh_cache-390bcf25-689d-46ad-bffb-3670c3729397" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.114615] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Releasing lock "refresh_cache-390bcf25-689d-46ad-bffb-3670c3729397" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.114985] env[62569]: DEBUG nova.compute.manager [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Instance network_info: |[{"id": "a466b80c-87f5-4bfd-ad2a-fc3610a634ed", "address": "fa:16:3e:38:93:3d", "network": {"id": "fa1d9d59-c210-4299-8da1-8441e1c2595e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-49047811", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.113", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ed0f5952cbd45eaa31512d6df93dc9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa466b80c-87", "ovs_interfaceid": "a466b80c-87f5-4bfd-ad2a-fc3610a634ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "ac9a07c1-60a1-4fd4-83c7-54ffc30240d8", "address": "fa:16:3e:d8:ba:94", "network": {"id": "d2c1889e-76c5-46d6-8d7c-918bf384a898", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1640224982", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.148", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "2ed0f5952cbd45eaa31512d6df93dc9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6e940e5-e083-4238-973e-f1b4e2a3a5c7", "external-id": "nsx-vlan-transportzone-64", "segmentation_id": 64, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac9a07c1-60", "ovs_interfaceid": "ac9a07c1-60a1-4fd4-83c7-54ffc30240d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 811.115304] env[62569]: DEBUG oslo_concurrency.lockutils [req-9813ccbf-358c-4a80-8eb0-6a2b780f6ef0 req-369592b3-87ef-49ed-8211-5daa3b38905b service nova] Acquired lock "refresh_cache-390bcf25-689d-46ad-bffb-3670c3729397" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.115493] env[62569]: DEBUG nova.network.neutron [req-9813ccbf-358c-4a80-8eb0-6a2b780f6ef0 req-369592b3-87ef-49ed-8211-5daa3b38905b service nova] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Refreshing network info cache for port ac9a07c1-60a1-4fd4-83c7-54ffc30240d8 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 811.116973] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:38:93:3d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '816c6e38-e200-4544-8c5b-9fc3e16c5761', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a466b80c-87f5-4bfd-ad2a-fc3610a634ed', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:ba:94', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd6e940e5-e083-4238-973e-f1b4e2a3a5c7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ac9a07c1-60a1-4fd4-83c7-54ffc30240d8', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 811.127531] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Creating folder: Project (2ed0f5952cbd45eaa31512d6df93dc9e). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 811.128721] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6c9fafb6-1e77-4a33-8942-e65f987ac764 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.140698] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Created folder: Project (2ed0f5952cbd45eaa31512d6df93dc9e) in parent group-v269330. [ 811.140984] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Creating folder: Instances. Parent ref: group-v269412. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 811.141260] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-df3b9abb-2315-43d7-8796-19fc880e2b54 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.150948] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Created folder: Instances in parent group-v269412. [ 811.150948] env[62569]: DEBUG oslo.service.loopingcall [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 811.150948] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 811.150948] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-49b0d347-dc76-4f06-9298-f06f78f042de {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.175062] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 811.175062] env[62569]: value = "task-1250097" [ 811.175062] env[62569]: _type = "Task" [ 811.175062] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.186261] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250097, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.226473] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250093, 'name': CreateVM_Task, 'duration_secs': 0.356102} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.226685] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 811.227408] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': 'df0f16e4-d160-4dca-94f7-563d020d5b2b', 'guest_format': None, 'device_type': None, 'disk_bus': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269356', 'volume_id': '3fb6b4c3-cf28-44c4-9606-7430ec94b7f6', 'name': 'volume-3fb6b4c3-cf28-44c4-9606-7430ec94b7f6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'fdeac8b8-a34d-4b4c-8f72-281f84e22ebe', 'attached_at': '', 'detached_at': '', 'volume_id': '3fb6b4c3-cf28-44c4-9606-7430ec94b7f6', 'serial': '3fb6b4c3-cf28-44c4-9606-7430ec94b7f6'}, 'mount_device': '/dev/sda', 'delete_on_termination': True, 'boot_index': 0, 'volume_type': None}], 'swap': None} {{(pid=62569) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 811.227680] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Root volume attach. Driver type: vmdk {{(pid=62569) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 811.228779] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01069a6a-5f19-46cd-9812-dd90597c1d0c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.239500] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c224ab88-8deb-44dd-8110-5b72f1a279e4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.248999] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f8e4069-f8f3-4b31-ba12-6f7dee00e700 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.261959] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-2f3214f0-bfd8-4ab8-a17b-854efbad123a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.269571] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Waiting for the task: (returnval){ [ 811.269571] env[62569]: value = "task-1250098" [ 811.269571] env[62569]: _type = "Task" [ 811.269571] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.279366] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250098, 'name': RelocateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.292900] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250094, 'name': Rename_Task, 'duration_secs': 0.15643} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.293233] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 811.293632] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-83727891-5463-4ec6-ae1c-fb2c3e7bb566 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.296030] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.296079] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.296561] env[62569]: DEBUG nova.network.neutron [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 811.303999] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 811.303999] env[62569]: value = "task-1250099" [ 811.303999] env[62569]: _type = "Task" [ 811.303999] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.317942] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250099, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.469265] env[62569]: DEBUG oslo_vmware.api [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250090, 'name': PowerOnVM_Task, 'duration_secs': 0.686403} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.472138] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 811.472385] env[62569]: INFO nova.compute.manager [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Took 8.79 seconds to spawn the instance on the hypervisor. [ 811.472575] env[62569]: DEBUG nova.compute.manager [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 811.474048] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe3f5cf5-7fbf-4c00-a54e-c56687b1daa6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.600909] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-337da6d8-305f-4ac1-ab2f-12e61f89f458 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.612482] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f4e6c79-a813-4bb0-91a5-85c1447dd5cf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.649847] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67e6e782-547d-4797-8556-0629499dd642 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.660457] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cbc8af0-bb62-4413-b749-c9d17a8fe5bd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.676806] env[62569]: DEBUG nova.compute.provider_tree [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.688866] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250097, 'name': CreateVM_Task, 'duration_secs': 0.417996} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.689074] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 811.690427] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.690669] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.691017] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 811.692178] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6e7dc62-4078-41d1-ac96-5a8749758b9d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.697857] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Acquiring lock "cefbca2e-609d-4954-bec6-52ffe095446f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.698184] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Lock "cefbca2e-609d-4954-bec6-52ffe095446f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.698413] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Acquiring lock "cefbca2e-609d-4954-bec6-52ffe095446f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.698633] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Lock "cefbca2e-609d-4954-bec6-52ffe095446f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.698816] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Lock "cefbca2e-609d-4954-bec6-52ffe095446f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.702679] env[62569]: INFO nova.compute.manager [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Terminating instance [ 811.707690] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Waiting for the task: (returnval){ [ 811.707690] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]525289ac-60f1-1552-9488-cf0a42c390de" [ 811.707690] env[62569]: _type = "Task" [ 811.707690] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.725076] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]525289ac-60f1-1552-9488-cf0a42c390de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.781769] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250098, 'name': RelocateVM_Task} progress is 40%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.793316] env[62569]: DEBUG nova.network.neutron [req-f198bf83-4ae2-4c1d-93e0-3ee8e9f00236 req-38f8cf75-16ab-479d-910d-25ae285b8d47 service nova] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Updated VIF entry in instance network info cache for port da0de7a0-922e-46b2-842b-a222e31de3b2. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 811.793913] env[62569]: DEBUG nova.network.neutron [req-f198bf83-4ae2-4c1d-93e0-3ee8e9f00236 req-38f8cf75-16ab-479d-910d-25ae285b8d47 service nova] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Updating instance_info_cache with network_info: [{"id": "da0de7a0-922e-46b2-842b-a222e31de3b2", "address": "fa:16:3e:eb:02:d2", "network": {"id": "ea8253a1-5878-4093-9acf-1426d966475b", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-616956732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3de498c4aa74ba698deedcacb9b44f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d33839ae-40ca-471b-92e3-eb282b920682", "external-id": "nsx-vlan-transportzone-416", "segmentation_id": 416, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda0de7a0-92", "ovs_interfaceid": "da0de7a0-922e-46b2-842b-a222e31de3b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.817852] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250099, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.860688] env[62569]: DEBUG nova.network.neutron [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 812.002234] env[62569]: INFO nova.compute.manager [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Took 32.25 seconds to build instance. [ 812.102239] env[62569]: DEBUG nova.network.neutron [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Updating instance_info_cache with network_info: [{"id": "90b3aaef-2e72-486f-be8d-a4202a6cb0c6", "address": "fa:16:3e:0c:8f:dd", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90b3aaef-2e", "ovs_interfaceid": "90b3aaef-2e72-486f-be8d-a4202a6cb0c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.105212] env[62569]: DEBUG nova.network.neutron [req-9813ccbf-358c-4a80-8eb0-6a2b780f6ef0 req-369592b3-87ef-49ed-8211-5daa3b38905b service nova] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Updated VIF entry in instance network info cache for port ac9a07c1-60a1-4fd4-83c7-54ffc30240d8. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 812.105676] env[62569]: DEBUG nova.network.neutron [req-9813ccbf-358c-4a80-8eb0-6a2b780f6ef0 req-369592b3-87ef-49ed-8211-5daa3b38905b service nova] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Updating instance_info_cache with network_info: [{"id": "a466b80c-87f5-4bfd-ad2a-fc3610a634ed", "address": "fa:16:3e:38:93:3d", "network": {"id": "fa1d9d59-c210-4299-8da1-8441e1c2595e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-49047811", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.113", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ed0f5952cbd45eaa31512d6df93dc9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "816c6e38-e200-4544-8c5b-9fc3e16c5761", "external-id": "nsx-vlan-transportzone-195", "segmentation_id": 195, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa466b80c-87", "ovs_interfaceid": "a466b80c-87f5-4bfd-ad2a-fc3610a634ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "ac9a07c1-60a1-4fd4-83c7-54ffc30240d8", "address": "fa:16:3e:d8:ba:94", "network": {"id": "d2c1889e-76c5-46d6-8d7c-918bf384a898", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1640224982", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.148", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "2ed0f5952cbd45eaa31512d6df93dc9e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6e940e5-e083-4238-973e-f1b4e2a3a5c7", "external-id": "nsx-vlan-transportzone-64", "segmentation_id": 64, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac9a07c1-60", "ovs_interfaceid": "ac9a07c1-60a1-4fd4-83c7-54ffc30240d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.180767] env[62569]: DEBUG nova.scheduler.client.report [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 812.214519] env[62569]: DEBUG nova.compute.manager [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 812.214764] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 812.215666] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67c9ca61-7931-4e62-8eb4-677c94e701d5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.223307] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]525289ac-60f1-1552-9488-cf0a42c390de, 'name': SearchDatastore_Task, 'duration_secs': 0.025371} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.223918] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.224199] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 812.224505] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.224614] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.224791] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 812.225097] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-23b0aa31-9c08-4346-9e8a-5533d35f3905 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.229349] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 812.229909] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c579734f-8f01-4229-8cc4-8efa36042225 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.236500] env[62569]: DEBUG oslo_vmware.api [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Waiting for the task: (returnval){ [ 812.236500] env[62569]: value = "task-1250100" [ 812.236500] env[62569]: _type = "Task" [ 812.236500] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.237879] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 812.238093] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 812.243434] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77ee8c17-765f-4aae-8b41-149ef0b707bb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.249953] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Waiting for the task: (returnval){ [ 812.249953] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]521fb829-8433-a982-85d9-f8ecf17bebac" [ 812.249953] env[62569]: _type = "Task" [ 812.249953] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.257101] env[62569]: DEBUG oslo_vmware.api [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Task: {'id': task-1250100, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.262134] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]521fb829-8433-a982-85d9-f8ecf17bebac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.282031] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250098, 'name': RelocateVM_Task} progress is 54%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.298981] env[62569]: DEBUG oslo_concurrency.lockutils [req-f198bf83-4ae2-4c1d-93e0-3ee8e9f00236 req-38f8cf75-16ab-479d-910d-25ae285b8d47 service nova] Releasing lock "refresh_cache-fdeac8b8-a34d-4b4c-8f72-281f84e22ebe" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.317709] env[62569]: DEBUG oslo_vmware.api [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250099, 'name': PowerOnVM_Task, 'duration_secs': 0.532974} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.318351] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 812.318351] env[62569]: DEBUG nova.compute.manager [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 812.319605] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e729fe23-b2cf-41cb-9f73-45cbd6d24c51 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.464918] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Acquiring lock "77a1b192-6aff-4fee-93d7-57cebcdce626" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.465090] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Lock "77a1b192-6aff-4fee-93d7-57cebcdce626" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.465282] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Acquiring lock "77a1b192-6aff-4fee-93d7-57cebcdce626-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.467281] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Lock "77a1b192-6aff-4fee-93d7-57cebcdce626-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.002s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.467492] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Lock "77a1b192-6aff-4fee-93d7-57cebcdce626-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.471585] env[62569]: INFO nova.compute.manager [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Terminating instance [ 812.506351] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dd438549-4cea-4584-83f5-907cf31d4d87 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "807d3025-d6a7-4778-a829-a61e2c7495c9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.939s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.605861] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Releasing lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.606268] env[62569]: DEBUG nova.compute.manager [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Instance network_info: |[{"id": "90b3aaef-2e72-486f-be8d-a4202a6cb0c6", "address": "fa:16:3e:0c:8f:dd", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90b3aaef-2e", "ovs_interfaceid": "90b3aaef-2e72-486f-be8d-a4202a6cb0c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 812.606885] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:8f:dd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b356db78-99c7-4464-822c-fc7e193f7878', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '90b3aaef-2e72-486f-be8d-a4202a6cb0c6', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 812.618088] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Creating folder: Project (00bdb4c3f88f4d61834f7e961629ef3a). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 812.618807] env[62569]: DEBUG oslo_concurrency.lockutils [req-9813ccbf-358c-4a80-8eb0-6a2b780f6ef0 req-369592b3-87ef-49ed-8211-5daa3b38905b service nova] Releasing lock "refresh_cache-390bcf25-689d-46ad-bffb-3670c3729397" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.619100] env[62569]: DEBUG nova.compute.manager [req-9813ccbf-358c-4a80-8eb0-6a2b780f6ef0 req-369592b3-87ef-49ed-8211-5daa3b38905b service nova] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Received event network-vif-plugged-90b3aaef-2e72-486f-be8d-a4202a6cb0c6 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 812.619384] env[62569]: DEBUG oslo_concurrency.lockutils [req-9813ccbf-358c-4a80-8eb0-6a2b780f6ef0 req-369592b3-87ef-49ed-8211-5daa3b38905b service nova] Acquiring lock "b47cd2d7-0cd2-41af-8ed1-a6dfca323516-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.619524] env[62569]: DEBUG oslo_concurrency.lockutils [req-9813ccbf-358c-4a80-8eb0-6a2b780f6ef0 req-369592b3-87ef-49ed-8211-5daa3b38905b service nova] Lock "b47cd2d7-0cd2-41af-8ed1-a6dfca323516-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.619705] env[62569]: DEBUG oslo_concurrency.lockutils [req-9813ccbf-358c-4a80-8eb0-6a2b780f6ef0 req-369592b3-87ef-49ed-8211-5daa3b38905b service nova] Lock "b47cd2d7-0cd2-41af-8ed1-a6dfca323516-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.619878] env[62569]: DEBUG nova.compute.manager [req-9813ccbf-358c-4a80-8eb0-6a2b780f6ef0 req-369592b3-87ef-49ed-8211-5daa3b38905b service nova] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] No waiting events found dispatching network-vif-plugged-90b3aaef-2e72-486f-be8d-a4202a6cb0c6 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 812.620059] env[62569]: WARNING nova.compute.manager [req-9813ccbf-358c-4a80-8eb0-6a2b780f6ef0 req-369592b3-87ef-49ed-8211-5daa3b38905b service nova] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Received unexpected event network-vif-plugged-90b3aaef-2e72-486f-be8d-a4202a6cb0c6 for instance with vm_state building and task_state spawning. [ 812.620236] env[62569]: DEBUG nova.compute.manager [req-9813ccbf-358c-4a80-8eb0-6a2b780f6ef0 req-369592b3-87ef-49ed-8211-5daa3b38905b service nova] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Received event network-changed-90b3aaef-2e72-486f-be8d-a4202a6cb0c6 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 812.620395] env[62569]: DEBUG nova.compute.manager [req-9813ccbf-358c-4a80-8eb0-6a2b780f6ef0 req-369592b3-87ef-49ed-8211-5daa3b38905b service nova] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Refreshing instance network info cache due to event network-changed-90b3aaef-2e72-486f-be8d-a4202a6cb0c6. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 812.620670] env[62569]: DEBUG oslo_concurrency.lockutils [req-9813ccbf-358c-4a80-8eb0-6a2b780f6ef0 req-369592b3-87ef-49ed-8211-5daa3b38905b service nova] Acquiring lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.620821] env[62569]: DEBUG oslo_concurrency.lockutils [req-9813ccbf-358c-4a80-8eb0-6a2b780f6ef0 req-369592b3-87ef-49ed-8211-5daa3b38905b service nova] Acquired lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.620987] env[62569]: DEBUG nova.network.neutron [req-9813ccbf-358c-4a80-8eb0-6a2b780f6ef0 req-369592b3-87ef-49ed-8211-5daa3b38905b service nova] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Refreshing network info cache for port 90b3aaef-2e72-486f-be8d-a4202a6cb0c6 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 812.623036] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-05a5b528-b494-40cd-a598-1b9f231205cd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.634268] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Created folder: Project (00bdb4c3f88f4d61834f7e961629ef3a) in parent group-v269330. [ 812.634490] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Creating folder: Instances. Parent ref: group-v269415. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 812.634825] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bb220480-b021-4d6e-b071-15645bd019a5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.643751] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Created folder: Instances in parent group-v269415. [ 812.644030] env[62569]: DEBUG oslo.service.loopingcall [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 812.644246] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 812.644467] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c46b1248-569f-48bc-8db0-1afbcd9d1aee {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.663734] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 812.663734] env[62569]: value = "task-1250103" [ 812.663734] env[62569]: _type = "Task" [ 812.663734] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.672058] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250103, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.687754] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.496s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.688332] env[62569]: DEBUG nova.compute.manager [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 812.691874] env[62569]: DEBUG oslo_concurrency.lockutils [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.046s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.692146] env[62569]: DEBUG nova.objects.instance [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Lazy-loading 'resources' on Instance uuid 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 812.751557] env[62569]: DEBUG oslo_vmware.api [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Task: {'id': task-1250100, 'name': PowerOffVM_Task, 'duration_secs': 0.40893} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.751557] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 812.751557] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 812.755216] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f1111d8f-ccfe-4a07-bd22-4c6939c596f9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.762562] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]521fb829-8433-a982-85d9-f8ecf17bebac, 'name': SearchDatastore_Task, 'duration_secs': 0.027498} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.763469] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f219eac-79b1-4bed-b6a6-d4e7a7916925 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.772114] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Waiting for the task: (returnval){ [ 812.772114] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529108bf-591c-18f5-2a00-0e3c062f334c" [ 812.772114] env[62569]: _type = "Task" [ 812.772114] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.785666] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250098, 'name': RelocateVM_Task} progress is 67%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.790104] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529108bf-591c-18f5-2a00-0e3c062f334c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.847962] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.975747] env[62569]: DEBUG nova.compute.manager [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 812.976040] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 812.978754] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab192d40-f0ad-4e02-bb68-a8209c8c2845 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.990402] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 812.990734] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3b0a0f82-54e3-4db4-8274-2b5895fd361b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.999091] env[62569]: DEBUG oslo_vmware.api [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Waiting for the task: (returnval){ [ 812.999091] env[62569]: value = "task-1250105" [ 812.999091] env[62569]: _type = "Task" [ 812.999091] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.010214] env[62569]: DEBUG nova.compute.manager [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 813.014532] env[62569]: DEBUG oslo_vmware.api [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Task: {'id': task-1250105, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.176837] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250103, 'name': CreateVM_Task} progress is 25%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.196381] env[62569]: DEBUG nova.compute.utils [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 813.205697] env[62569]: DEBUG nova.compute.manager [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 813.205697] env[62569]: DEBUG nova.network.neutron [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 813.298359] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529108bf-591c-18f5-2a00-0e3c062f334c, 'name': SearchDatastore_Task, 'duration_secs': 0.018376} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.300781] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.301957] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 390bcf25-689d-46ad-bffb-3670c3729397/390bcf25-689d-46ad-bffb-3670c3729397.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 813.301957] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250098, 'name': RelocateVM_Task} progress is 81%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.301957] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4fcf929d-4b07-4762-99fd-2d8d4ce093d9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.315667] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Waiting for the task: (returnval){ [ 813.315667] env[62569]: value = "task-1250106" [ 813.315667] env[62569]: _type = "Task" [ 813.315667] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.329794] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Task: {'id': task-1250106, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.351556] env[62569]: DEBUG nova.policy [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '50780a0da3b14122aa6287c1a482f671', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7f7fb3e698324f8e851aefbee74ac669', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 813.511196] env[62569]: DEBUG oslo_vmware.api [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Task: {'id': task-1250105, 'name': PowerOffVM_Task, 'duration_secs': 0.350034} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.514437] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 813.514559] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 813.520022] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4157822d-0c7f-43b3-9458-b0e7732eb7b4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.536035] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.633684] env[62569]: DEBUG nova.compute.manager [req-95fd3955-12c8-4c57-bdbe-d11a1b8d2f04 req-c96072f2-6300-41ab-8e56-d174bdbd7ad1 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Received event network-changed-9d855102-33ab-4574-b791-dfa6fb5b0f09 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 813.633986] env[62569]: DEBUG nova.compute.manager [req-95fd3955-12c8-4c57-bdbe-d11a1b8d2f04 req-c96072f2-6300-41ab-8e56-d174bdbd7ad1 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Refreshing instance network info cache due to event network-changed-9d855102-33ab-4574-b791-dfa6fb5b0f09. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 813.634233] env[62569]: DEBUG oslo_concurrency.lockutils [req-95fd3955-12c8-4c57-bdbe-d11a1b8d2f04 req-c96072f2-6300-41ab-8e56-d174bdbd7ad1 service nova] Acquiring lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.634381] env[62569]: DEBUG oslo_concurrency.lockutils [req-95fd3955-12c8-4c57-bdbe-d11a1b8d2f04 req-c96072f2-6300-41ab-8e56-d174bdbd7ad1 service nova] Acquired lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.634549] env[62569]: DEBUG nova.network.neutron [req-95fd3955-12c8-4c57-bdbe-d11a1b8d2f04 req-c96072f2-6300-41ab-8e56-d174bdbd7ad1 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Refreshing network info cache for port 9d855102-33ab-4574-b791-dfa6fb5b0f09 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 813.642880] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d872ed2e-7d5c-4046-9884-3ca60c095264 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.654939] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb07d433-3aad-49a7-9204-eefe5d974b8e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.695317] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-010d3aec-f947-4fb1-8edd-b73c9349707d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.699941] env[62569]: DEBUG nova.network.neutron [req-9813ccbf-358c-4a80-8eb0-6a2b780f6ef0 req-369592b3-87ef-49ed-8211-5daa3b38905b service nova] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Updated VIF entry in instance network info cache for port 90b3aaef-2e72-486f-be8d-a4202a6cb0c6. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 813.700353] env[62569]: DEBUG nova.network.neutron [req-9813ccbf-358c-4a80-8eb0-6a2b780f6ef0 req-369592b3-87ef-49ed-8211-5daa3b38905b service nova] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Updating instance_info_cache with network_info: [{"id": "90b3aaef-2e72-486f-be8d-a4202a6cb0c6", "address": "fa:16:3e:0c:8f:dd", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90b3aaef-2e", "ovs_interfaceid": "90b3aaef-2e72-486f-be8d-a4202a6cb0c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.708926] env[62569]: DEBUG nova.compute.manager [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 813.715042] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250103, 'name': CreateVM_Task} progress is 25%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.728933] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-840fac86-bbab-494d-a51a-324ff796964d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.748927] env[62569]: DEBUG nova.compute.provider_tree [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.781694] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250098, 'name': RelocateVM_Task} progress is 95%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.828347] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Task: {'id': task-1250106, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.972126] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 813.972416] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 813.973097] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Deleting the datastore file [datastore1] cefbca2e-609d-4954-bec6-52ffe095446f {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 813.973580] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9e2d4919-00b8-4659-b63c-2aceb4022585 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.977107] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 813.977550] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 813.977995] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Deleting the datastore file [datastore1] 77a1b192-6aff-4fee-93d7-57cebcdce626 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 813.978347] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f4721aa9-2bd9-4f6e-b383-3a4f67afaeb5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.986034] env[62569]: DEBUG oslo_vmware.api [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Waiting for the task: (returnval){ [ 813.986034] env[62569]: value = "task-1250109" [ 813.986034] env[62569]: _type = "Task" [ 813.986034] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.987286] env[62569]: DEBUG oslo_vmware.api [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Waiting for the task: (returnval){ [ 813.987286] env[62569]: value = "task-1250108" [ 813.987286] env[62569]: _type = "Task" [ 813.987286] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.999825] env[62569]: DEBUG oslo_vmware.api [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Task: {'id': task-1250109, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.005934] env[62569]: DEBUG oslo_vmware.api [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Task: {'id': task-1250108, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.019582] env[62569]: DEBUG nova.network.neutron [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Successfully created port: cbc4229a-f42d-463c-9a46-6a4d6b3581b0 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 814.204948] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250103, 'name': CreateVM_Task} progress is 25%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.211092] env[62569]: DEBUG oslo_concurrency.lockutils [req-9813ccbf-358c-4a80-8eb0-6a2b780f6ef0 req-369592b3-87ef-49ed-8211-5daa3b38905b service nova] Releasing lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.252830] env[62569]: DEBUG nova.scheduler.client.report [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 814.290679] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250098, 'name': RelocateVM_Task} progress is 97%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.328367] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Task: {'id': task-1250106, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.941685} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.328909] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 390bcf25-689d-46ad-bffb-3670c3729397/390bcf25-689d-46ad-bffb-3670c3729397.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 814.333018] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 814.333018] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f4607867-6616-4a23-b666-7e307efd1b5a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.336189] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Waiting for the task: (returnval){ [ 814.336189] env[62569]: value = "task-1250110" [ 814.336189] env[62569]: _type = "Task" [ 814.336189] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.347205] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Task: {'id': task-1250110, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.501140] env[62569]: DEBUG oslo_vmware.api [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Task: {'id': task-1250109, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.504282] env[62569]: DEBUG oslo_vmware.api [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Task: {'id': task-1250108, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.559304] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "eb70341b-4282-4eca-b6a2-374db7c521c5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.559568] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "eb70341b-4282-4eca-b6a2-374db7c521c5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.559931] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "eb70341b-4282-4eca-b6a2-374db7c521c5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.563017] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "eb70341b-4282-4eca-b6a2-374db7c521c5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.563017] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "eb70341b-4282-4eca-b6a2-374db7c521c5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.563017] env[62569]: INFO nova.compute.manager [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Terminating instance [ 814.596101] env[62569]: DEBUG nova.network.neutron [req-95fd3955-12c8-4c57-bdbe-d11a1b8d2f04 req-c96072f2-6300-41ab-8e56-d174bdbd7ad1 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Updated VIF entry in instance network info cache for port 9d855102-33ab-4574-b791-dfa6fb5b0f09. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 814.596638] env[62569]: DEBUG nova.network.neutron [req-95fd3955-12c8-4c57-bdbe-d11a1b8d2f04 req-c96072f2-6300-41ab-8e56-d174bdbd7ad1 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Updating instance_info_cache with network_info: [{"id": "9d855102-33ab-4574-b791-dfa6fb5b0f09", "address": "fa:16:3e:b3:4b:f5", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d855102-33", "ovs_interfaceid": "9d855102-33ab-4574-b791-dfa6fb5b0f09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.703513] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250103, 'name': CreateVM_Task, 'duration_secs': 1.881184} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.703513] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 814.704014] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.704204] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.704539] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 814.704808] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a525855-59b4-4b1a-8d6d-c2cf8e75778b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.709969] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 814.709969] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520b6d6a-b8f2-3f3d-9569-21a630598cc4" [ 814.709969] env[62569]: _type = "Task" [ 814.709969] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.718702] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520b6d6a-b8f2-3f3d-9569-21a630598cc4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.730127] env[62569]: DEBUG nova.compute.manager [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 814.760069] env[62569]: DEBUG nova.virt.hardware [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 814.760347] env[62569]: DEBUG nova.virt.hardware [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 814.760505] env[62569]: DEBUG nova.virt.hardware [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 814.761376] env[62569]: DEBUG nova.virt.hardware [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 814.761376] env[62569]: DEBUG nova.virt.hardware [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 814.761376] env[62569]: DEBUG nova.virt.hardware [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 814.761376] env[62569]: DEBUG nova.virt.hardware [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 814.761827] env[62569]: DEBUG nova.virt.hardware [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 814.761827] env[62569]: DEBUG nova.virt.hardware [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 814.761888] env[62569]: DEBUG nova.virt.hardware [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 814.762108] env[62569]: DEBUG nova.virt.hardware [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 814.764091] env[62569]: DEBUG oslo_concurrency.lockutils [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.071s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.765860] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67705ead-0cbf-42f3-9359-59a417ba2a3e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.770093] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.341s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.772187] env[62569]: INFO nova.compute.claims [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 814.783830] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51cfb87b-cf1b-40bc-84f1-5115337fb96b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.791398] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250098, 'name': RelocateVM_Task} progress is 97%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.805217] env[62569]: INFO nova.scheduler.client.report [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Deleted allocations for instance 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010 [ 814.847177] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Task: {'id': task-1250110, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084987} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.847497] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 814.848354] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee5497dc-4678-4851-9348-c5ae05bbd671 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.881104] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] 390bcf25-689d-46ad-bffb-3670c3729397/390bcf25-689d-46ad-bffb-3670c3729397.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 814.881104] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64b2f37e-921d-4e33-83ce-75522a7fd007 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.902859] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Waiting for the task: (returnval){ [ 814.902859] env[62569]: value = "task-1250111" [ 814.902859] env[62569]: _type = "Task" [ 814.902859] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.911696] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Task: {'id': task-1250111, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.000795] env[62569]: DEBUG oslo_vmware.api [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Task: {'id': task-1250109, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.551052} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.003949] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 815.004171] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 815.004364] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 815.004562] env[62569]: INFO nova.compute.manager [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Took 2.03 seconds to destroy the instance on the hypervisor. [ 815.004813] env[62569]: DEBUG oslo.service.loopingcall [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 815.005039] env[62569]: DEBUG oslo_vmware.api [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Task: {'id': task-1250108, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.554045} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.005251] env[62569]: DEBUG nova.compute.manager [-] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 815.005349] env[62569]: DEBUG nova.network.neutron [-] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 815.007069] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 815.007274] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 815.007460] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 815.007635] env[62569]: INFO nova.compute.manager [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Took 2.79 seconds to destroy the instance on the hypervisor. [ 815.007863] env[62569]: DEBUG oslo.service.loopingcall [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 815.008078] env[62569]: DEBUG nova.compute.manager [-] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 815.008173] env[62569]: DEBUG nova.network.neutron [-] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 815.066609] env[62569]: DEBUG nova.compute.manager [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 815.067032] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 815.068060] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bfb3f72-5446-42d4-955b-28383e87660c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.075992] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 815.076277] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f0a566e9-1bf8-476f-9c59-30b6f87ce441 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.082307] env[62569]: DEBUG oslo_vmware.api [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 815.082307] env[62569]: value = "task-1250112" [ 815.082307] env[62569]: _type = "Task" [ 815.082307] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.095044] env[62569]: DEBUG oslo_vmware.api [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250112, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.098843] env[62569]: DEBUG oslo_concurrency.lockutils [req-95fd3955-12c8-4c57-bdbe-d11a1b8d2f04 req-c96072f2-6300-41ab-8e56-d174bdbd7ad1 service nova] Releasing lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.221801] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520b6d6a-b8f2-3f3d-9569-21a630598cc4, 'name': SearchDatastore_Task, 'duration_secs': 0.024433} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.222191] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.222495] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 815.222751] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.223018] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.223226] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 815.223509] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9f690ec1-4d8c-4449-8043-7207123a8b5b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.232680] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 815.232740] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 815.233766] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-606d3d42-489c-4236-bc11-37455cd4bfc9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.240429] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 815.240429] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52369500-28dc-eb63-2f92-60891b078578" [ 815.240429] env[62569]: _type = "Task" [ 815.240429] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.250837] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52369500-28dc-eb63-2f92-60891b078578, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.254275] env[62569]: DEBUG oslo_concurrency.lockutils [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "31ce29fa-4f60-4404-b830-21ad196f78b5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.254548] env[62569]: DEBUG oslo_concurrency.lockutils [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "31ce29fa-4f60-4404-b830-21ad196f78b5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.290952] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250098, 'name': RelocateVM_Task} progress is 98%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.313723] env[62569]: DEBUG oslo_concurrency.lockutils [None req-790d461e-b887-40a3-aee9-9ff8f6dbf858 tempest-ServersNegativeTestMultiTenantJSON-1526652638 tempest-ServersNegativeTestMultiTenantJSON-1526652638-project-member] Lock "0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.192s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.412221] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Task: {'id': task-1250111, 'name': ReconfigVM_Task, 'duration_secs': 0.305796} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.412844] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Reconfigured VM instance instance-00000045 to attach disk [datastore1] 390bcf25-689d-46ad-bffb-3670c3729397/390bcf25-689d-46ad-bffb-3670c3729397.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 815.413658] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fa1c9197-e903-4471-93f4-d13f9c2f55fb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.422585] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Waiting for the task: (returnval){ [ 815.422585] env[62569]: value = "task-1250113" [ 815.422585] env[62569]: _type = "Task" [ 815.422585] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.430376] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Task: {'id': task-1250113, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.593633] env[62569]: DEBUG oslo_vmware.api [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250112, 'name': PowerOffVM_Task, 'duration_secs': 0.227953} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.593919] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 815.594323] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 815.594616] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c94980fe-8a8d-46ea-9c4f-bcc339ca5620 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.665420] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 815.665784] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 815.665987] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Deleting the datastore file [datastore1] eb70341b-4282-4eca-b6a2-374db7c521c5 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 815.667330] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2a0d112f-1437-4e51-af03-0a7382f2b054 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.671240] env[62569]: DEBUG nova.compute.manager [req-3e832d80-2b7d-462d-8f3b-c057d1f79b30 req-dd20366a-ece0-464c-8d44-d4c340b41209 service nova] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Received event network-vif-deleted-8c024324-0c17-4626-984c-a4e3309e1b44 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 815.671441] env[62569]: INFO nova.compute.manager [req-3e832d80-2b7d-462d-8f3b-c057d1f79b30 req-dd20366a-ece0-464c-8d44-d4c340b41209 service nova] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Neutron deleted interface 8c024324-0c17-4626-984c-a4e3309e1b44; detaching it from the instance and deleting it from the info cache [ 815.671638] env[62569]: DEBUG nova.network.neutron [req-3e832d80-2b7d-462d-8f3b-c057d1f79b30 req-dd20366a-ece0-464c-8d44-d4c340b41209 service nova] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.681208] env[62569]: DEBUG oslo_vmware.api [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 815.681208] env[62569]: value = "task-1250115" [ 815.681208] env[62569]: _type = "Task" [ 815.681208] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.694916] env[62569]: DEBUG oslo_vmware.api [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250115, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.751159] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52369500-28dc-eb63-2f92-60891b078578, 'name': SearchDatastore_Task, 'duration_secs': 0.034599} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.752065] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30e1639c-5500-40ec-a0bb-d72cd36688da {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.757929] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 815.757929] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5268798a-20e8-4704-64ad-58f3de5a6962" [ 815.757929] env[62569]: _type = "Task" [ 815.757929] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.767026] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5268798a-20e8-4704-64ad-58f3de5a6962, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.796802] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250098, 'name': RelocateVM_Task} progress is 98%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.865270] env[62569]: DEBUG nova.compute.manager [req-eda6a49d-63f3-4577-8b2b-3db48a7d1c00 req-89beffa5-c53c-49d4-af9f-5b6ba015d25c service nova] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Received event network-vif-plugged-cbc4229a-f42d-463c-9a46-6a4d6b3581b0 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 815.865270] env[62569]: DEBUG oslo_concurrency.lockutils [req-eda6a49d-63f3-4577-8b2b-3db48a7d1c00 req-89beffa5-c53c-49d4-af9f-5b6ba015d25c service nova] Acquiring lock "cb9d8991-b7ba-4bf7-9d22-1391f4cea768-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.865270] env[62569]: DEBUG oslo_concurrency.lockutils [req-eda6a49d-63f3-4577-8b2b-3db48a7d1c00 req-89beffa5-c53c-49d4-af9f-5b6ba015d25c service nova] Lock "cb9d8991-b7ba-4bf7-9d22-1391f4cea768-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.865270] env[62569]: DEBUG oslo_concurrency.lockutils [req-eda6a49d-63f3-4577-8b2b-3db48a7d1c00 req-89beffa5-c53c-49d4-af9f-5b6ba015d25c service nova] Lock "cb9d8991-b7ba-4bf7-9d22-1391f4cea768-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.865270] env[62569]: DEBUG nova.compute.manager [req-eda6a49d-63f3-4577-8b2b-3db48a7d1c00 req-89beffa5-c53c-49d4-af9f-5b6ba015d25c service nova] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] No waiting events found dispatching network-vif-plugged-cbc4229a-f42d-463c-9a46-6a4d6b3581b0 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 815.865270] env[62569]: WARNING nova.compute.manager [req-eda6a49d-63f3-4577-8b2b-3db48a7d1c00 req-89beffa5-c53c-49d4-af9f-5b6ba015d25c service nova] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Received unexpected event network-vif-plugged-cbc4229a-f42d-463c-9a46-6a4d6b3581b0 for instance with vm_state building and task_state spawning. [ 815.934810] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Task: {'id': task-1250113, 'name': Rename_Task, 'duration_secs': 0.144431} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.938630] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 815.939995] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-00e249f6-239b-405d-b767-098b402a337c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.948303] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Waiting for the task: (returnval){ [ 815.948303] env[62569]: value = "task-1250116" [ 815.948303] env[62569]: _type = "Task" [ 815.948303] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.961142] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Task: {'id': task-1250116, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.961513] env[62569]: DEBUG nova.network.neutron [-] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.013676] env[62569]: DEBUG nova.compute.manager [req-7117d4d9-54e6-4448-a784-34e33489085a req-11687f01-21de-4c60-b64d-055fb7523cef service nova] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Received event network-vif-deleted-6a7245e5-dc62-4799-b00a-34cf50cd5a74 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 816.013676] env[62569]: INFO nova.compute.manager [req-7117d4d9-54e6-4448-a784-34e33489085a req-11687f01-21de-4c60-b64d-055fb7523cef service nova] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Neutron deleted interface 6a7245e5-dc62-4799-b00a-34cf50cd5a74; detaching it from the instance and deleting it from the info cache [ 816.014103] env[62569]: DEBUG nova.network.neutron [req-7117d4d9-54e6-4448-a784-34e33489085a req-11687f01-21de-4c60-b64d-055fb7523cef service nova] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.174406] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1c5bf7e5-87f0-40f5-b490-f42119472a79 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.183303] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de4d4d4-12bb-4bdf-aaca-22b159adbeb5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.195910] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a9d7cdf-3555-46a4-ae7d-92bf5b18c3d4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.210998] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-711e0560-89b9-480d-af12-279299ad5216 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.217854] env[62569]: DEBUG oslo_vmware.api [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250115, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.333258} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.228751] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 816.228999] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 816.229215] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 816.229490] env[62569]: INFO nova.compute.manager [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Took 1.16 seconds to destroy the instance on the hypervisor. [ 816.229739] env[62569]: DEBUG oslo.service.loopingcall [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.229979] env[62569]: DEBUG nova.compute.manager [req-3e832d80-2b7d-462d-8f3b-c057d1f79b30 req-dd20366a-ece0-464c-8d44-d4c340b41209 service nova] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Detach interface failed, port_id=8c024324-0c17-4626-984c-a4e3309e1b44, reason: Instance cefbca2e-609d-4954-bec6-52ffe095446f could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 816.230947] env[62569]: DEBUG nova.compute.manager [-] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 816.231067] env[62569]: DEBUG nova.network.neutron [-] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 816.265291] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fead53a-5d1a-4bd7-bb31-6cf816c93b20 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.280596] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b2818a2-d4b4-48c0-8ede-c939a910acd7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.287611] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5268798a-20e8-4704-64ad-58f3de5a6962, 'name': SearchDatastore_Task, 'duration_secs': 0.01153} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.287998] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.288285] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] b47cd2d7-0cd2-41af-8ed1-a6dfca323516/b47cd2d7-0cd2-41af-8ed1-a6dfca323516.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 816.292284] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-62a5ff30-9118-4978-bb45-03d3edd4da55 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.302682] env[62569]: DEBUG nova.compute.provider_tree [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 816.312593] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 816.312593] env[62569]: value = "task-1250117" [ 816.312593] env[62569]: _type = "Task" [ 816.312593] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.312873] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250098, 'name': RelocateVM_Task, 'duration_secs': 5.038917} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.313821] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Volume attach. Driver type: vmdk {{(pid=62569) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 816.314052] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269356', 'volume_id': '3fb6b4c3-cf28-44c4-9606-7430ec94b7f6', 'name': 'volume-3fb6b4c3-cf28-44c4-9606-7430ec94b7f6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'fdeac8b8-a34d-4b4c-8f72-281f84e22ebe', 'attached_at': '', 'detached_at': '', 'volume_id': '3fb6b4c3-cf28-44c4-9606-7430ec94b7f6', 'serial': '3fb6b4c3-cf28-44c4-9606-7430ec94b7f6'} {{(pid=62569) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 816.317576] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dd48575-e562-4e74-a8f2-74ef1469a745 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.325436] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250117, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.340049] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09b87bb1-e18e-41c6-a786-f3766ca38715 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.365263] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] volume-3fb6b4c3-cf28-44c4-9606-7430ec94b7f6/volume-3fb6b4c3-cf28-44c4-9606-7430ec94b7f6.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 816.365890] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6bb63721-12a9-43c4-ba81-5e7f4670e615 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.386154] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Waiting for the task: (returnval){ [ 816.386154] env[62569]: value = "task-1250118" [ 816.386154] env[62569]: _type = "Task" [ 816.386154] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.396149] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250118, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.426264] env[62569]: DEBUG nova.network.neutron [-] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.464594] env[62569]: DEBUG oslo_vmware.api [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Task: {'id': task-1250116, 'name': PowerOnVM_Task, 'duration_secs': 0.479734} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.464594] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 816.464594] env[62569]: INFO nova.compute.manager [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Took 11.34 seconds to spawn the instance on the hypervisor. [ 816.464594] env[62569]: DEBUG nova.compute.manager [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 816.464594] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eae643c-3dfa-40ca-8ee2-3ea91bf6f028 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.465863] env[62569]: INFO nova.compute.manager [-] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Took 1.46 seconds to deallocate network for instance. [ 816.485941] env[62569]: DEBUG nova.network.neutron [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Successfully updated port: cbc4229a-f42d-463c-9a46-6a4d6b3581b0 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 816.523178] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6efc0bca-11e3-4fec-9ac8-bf6a7eacb45c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.535888] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9a8fc04-5af4-408e-b647-75dbe29cff73 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.565597] env[62569]: DEBUG nova.compute.manager [req-7117d4d9-54e6-4448-a784-34e33489085a req-11687f01-21de-4c60-b64d-055fb7523cef service nova] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Detach interface failed, port_id=6a7245e5-dc62-4799-b00a-34cf50cd5a74, reason: Instance 77a1b192-6aff-4fee-93d7-57cebcdce626 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 816.812059] env[62569]: DEBUG nova.scheduler.client.report [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 816.828259] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250117, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.896634] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250118, 'name': ReconfigVM_Task, 'duration_secs': 0.300749} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.896944] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Reconfigured VM instance instance-00000046 to attach disk [datastore1] volume-3fb6b4c3-cf28-44c4-9606-7430ec94b7f6/volume-3fb6b4c3-cf28-44c4-9606-7430ec94b7f6.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 816.903351] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a631219-e201-4103-b5cc-13b826d7d78a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.919146] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Waiting for the task: (returnval){ [ 816.919146] env[62569]: value = "task-1250119" [ 816.919146] env[62569]: _type = "Task" [ 816.919146] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.927992] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250119, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.928609] env[62569]: INFO nova.compute.manager [-] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Took 1.92 seconds to deallocate network for instance. [ 816.977296] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.984172] env[62569]: INFO nova.compute.manager [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Took 34.83 seconds to build instance. [ 816.993624] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "refresh_cache-cb9d8991-b7ba-4bf7-9d22-1391f4cea768" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.993773] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquired lock "refresh_cache-cb9d8991-b7ba-4bf7-9d22-1391f4cea768" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.993925] env[62569]: DEBUG nova.network.neutron [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 817.086879] env[62569]: DEBUG nova.network.neutron [-] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.322457] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.553s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.323106] env[62569]: DEBUG nova.compute.manager [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 817.326162] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.023s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.327716] env[62569]: INFO nova.compute.claims [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 817.336697] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250117, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.586986} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.337017] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] b47cd2d7-0cd2-41af-8ed1-a6dfca323516/b47cd2d7-0cd2-41af-8ed1-a6dfca323516.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 817.337283] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 817.337586] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-151bb6a7-3f70-4aca-9936-d11ae8e59765 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.345643] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 817.345643] env[62569]: value = "task-1250120" [ 817.345643] env[62569]: _type = "Task" [ 817.345643] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.353754] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250120, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.429867] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250119, 'name': ReconfigVM_Task, 'duration_secs': 0.261018} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.430201] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269356', 'volume_id': '3fb6b4c3-cf28-44c4-9606-7430ec94b7f6', 'name': 'volume-3fb6b4c3-cf28-44c4-9606-7430ec94b7f6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'fdeac8b8-a34d-4b4c-8f72-281f84e22ebe', 'attached_at': '', 'detached_at': '', 'volume_id': '3fb6b4c3-cf28-44c4-9606-7430ec94b7f6', 'serial': '3fb6b4c3-cf28-44c4-9606-7430ec94b7f6'} {{(pid=62569) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 817.430825] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-23a4ed6c-e64e-4e7c-aa0a-08ea652f9c3a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.434707] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.437478] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Waiting for the task: (returnval){ [ 817.437478] env[62569]: value = "task-1250121" [ 817.437478] env[62569]: _type = "Task" [ 817.437478] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.445762] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250121, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.486961] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68ed64a9-fabc-4572-957b-5118c243cf29 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Lock "390bcf25-689d-46ad-bffb-3670c3729397" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.163s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.528650] env[62569]: DEBUG nova.network.neutron [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 817.590021] env[62569]: INFO nova.compute.manager [-] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Took 1.36 seconds to deallocate network for instance. [ 817.671190] env[62569]: DEBUG nova.network.neutron [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Updating instance_info_cache with network_info: [{"id": "cbc4229a-f42d-463c-9a46-6a4d6b3581b0", "address": "fa:16:3e:14:62:67", "network": {"id": "334a2e22-0249-48e3-a2f5-ac37df9771d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1048653043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f7fb3e698324f8e851aefbee74ac669", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f880ac2e-d532-4f54-87bb-998a8d1bca78", "external-id": "nsx-vlan-transportzone-491", "segmentation_id": 491, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcbc4229a-f4", "ovs_interfaceid": "cbc4229a-f42d-463c-9a46-6a4d6b3581b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.833325] env[62569]: DEBUG nova.compute.utils [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 817.836622] env[62569]: DEBUG nova.compute.manager [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 817.836784] env[62569]: DEBUG nova.network.neutron [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 817.860016] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250120, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.219046} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.860016] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 817.860016] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ce3a63-b809-4aed-a650-6ada2d0ef7ea {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.882885] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] b47cd2d7-0cd2-41af-8ed1-a6dfca323516/b47cd2d7-0cd2-41af-8ed1-a6dfca323516.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 817.883593] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c35b2d08-7cf4-4e11-9c42-1e65bfd0eb88 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.908665] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 817.908665] env[62569]: value = "task-1250122" [ 817.908665] env[62569]: _type = "Task" [ 817.908665] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.917977] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250122, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.946167] env[62569]: DEBUG nova.policy [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ccb9b3f4e38e4e81819d22e781661015', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7f988acdb9cd4ca28ca8916c2ede2db5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 817.959094] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250121, 'name': Rename_Task, 'duration_secs': 0.123577} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.959329] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 817.959521] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6c524ec9-e9d9-4a2a-ba6b-21377059be5e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.966038] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Waiting for the task: (returnval){ [ 817.966038] env[62569]: value = "task-1250123" [ 817.966038] env[62569]: _type = "Task" [ 817.966038] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.975893] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250123, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.989891] env[62569]: DEBUG nova.compute.manager [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 818.096579] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.109350] env[62569]: DEBUG nova.compute.manager [req-89f38a72-656c-47fb-a474-3bedfa0c4ea8 req-b903f194-d491-4ade-bd47-05984a8a2261 service nova] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Received event network-changed-cbc4229a-f42d-463c-9a46-6a4d6b3581b0 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 818.109563] env[62569]: DEBUG nova.compute.manager [req-89f38a72-656c-47fb-a474-3bedfa0c4ea8 req-b903f194-d491-4ade-bd47-05984a8a2261 service nova] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Refreshing instance network info cache due to event network-changed-cbc4229a-f42d-463c-9a46-6a4d6b3581b0. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 818.109759] env[62569]: DEBUG oslo_concurrency.lockutils [req-89f38a72-656c-47fb-a474-3bedfa0c4ea8 req-b903f194-d491-4ade-bd47-05984a8a2261 service nova] Acquiring lock "refresh_cache-cb9d8991-b7ba-4bf7-9d22-1391f4cea768" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.174046] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Releasing lock "refresh_cache-cb9d8991-b7ba-4bf7-9d22-1391f4cea768" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.174640] env[62569]: DEBUG nova.compute.manager [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Instance network_info: |[{"id": "cbc4229a-f42d-463c-9a46-6a4d6b3581b0", "address": "fa:16:3e:14:62:67", "network": {"id": "334a2e22-0249-48e3-a2f5-ac37df9771d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1048653043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f7fb3e698324f8e851aefbee74ac669", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f880ac2e-d532-4f54-87bb-998a8d1bca78", "external-id": "nsx-vlan-transportzone-491", "segmentation_id": 491, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcbc4229a-f4", "ovs_interfaceid": "cbc4229a-f42d-463c-9a46-6a4d6b3581b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 818.175106] env[62569]: DEBUG oslo_concurrency.lockutils [req-89f38a72-656c-47fb-a474-3bedfa0c4ea8 req-b903f194-d491-4ade-bd47-05984a8a2261 service nova] Acquired lock "refresh_cache-cb9d8991-b7ba-4bf7-9d22-1391f4cea768" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.175650] env[62569]: DEBUG nova.network.neutron [req-89f38a72-656c-47fb-a474-3bedfa0c4ea8 req-b903f194-d491-4ade-bd47-05984a8a2261 service nova] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Refreshing network info cache for port cbc4229a-f42d-463c-9a46-6a4d6b3581b0 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 818.179086] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:62:67', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f880ac2e-d532-4f54-87bb-998a8d1bca78', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cbc4229a-f42d-463c-9a46-6a4d6b3581b0', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 818.187519] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Creating folder: Project (7f7fb3e698324f8e851aefbee74ac669). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 818.188700] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3a16cf62-23f8-48c5-b80f-1ec73e1d4aa4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.200856] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Created folder: Project (7f7fb3e698324f8e851aefbee74ac669) in parent group-v269330. [ 818.201163] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Creating folder: Instances. Parent ref: group-v269418. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 818.201431] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a4d47960-6d70-4aa7-ad46-8e664ed28a45 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.210894] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Created folder: Instances in parent group-v269418. [ 818.211157] env[62569]: DEBUG oslo.service.loopingcall [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 818.211357] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 818.211567] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bd97814c-4e08-4202-9061-63cac353aaa5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.232480] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 818.232480] env[62569]: value = "task-1250126" [ 818.232480] env[62569]: _type = "Task" [ 818.232480] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.241006] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250126, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.297574] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Acquiring lock "390bcf25-689d-46ad-bffb-3670c3729397" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.298052] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Lock "390bcf25-689d-46ad-bffb-3670c3729397" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.298368] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Acquiring lock "390bcf25-689d-46ad-bffb-3670c3729397-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.298722] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Lock "390bcf25-689d-46ad-bffb-3670c3729397-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.298974] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Lock "390bcf25-689d-46ad-bffb-3670c3729397-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.301713] env[62569]: INFO nova.compute.manager [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Terminating instance [ 818.337840] env[62569]: DEBUG nova.compute.manager [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 818.419093] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250122, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.479202] env[62569]: DEBUG oslo_vmware.api [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250123, 'name': PowerOnVM_Task, 'duration_secs': 0.503779} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.482539] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 818.482783] env[62569]: INFO nova.compute.manager [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Took 9.10 seconds to spawn the instance on the hypervisor. [ 818.482972] env[62569]: DEBUG nova.compute.manager [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 818.484299] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-472f7489-7321-4c27-b6e7-a148c68303c0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.507048] env[62569]: DEBUG nova.network.neutron [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Successfully created port: 39e71a4f-3b81-4b34-8cc2-2a5352eca4ce {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 818.518133] env[62569]: DEBUG oslo_concurrency.lockutils [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.709345] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e5bee5a-ecba-4a4f-a55a-1f0083cf391c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.717078] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfaf811b-6e37-48b3-b4da-7194a197c362 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.766262] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47b76dc8-2e77-4497-a317-1ec17b9d0165 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.777486] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b20cfc27-cb83-40bb-b4be-f860153d7924 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.781303] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250126, 'name': CreateVM_Task, 'duration_secs': 0.332137} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.781485] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 818.782517] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.782686] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.783016] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 818.783265] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f20c26a4-44e2-4d93-b5a6-6ba53eb5923d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.793250] env[62569]: DEBUG nova.compute.provider_tree [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.798038] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 818.798038] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5213d248-db04-100c-0747-6658850b8960" [ 818.798038] env[62569]: _type = "Task" [ 818.798038] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.807080] env[62569]: DEBUG nova.compute.manager [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 818.807316] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 818.807829] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5213d248-db04-100c-0747-6658850b8960, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.808560] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1e8b11c-ae47-46fe-9368-790929c70d06 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.815589] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 818.816429] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b77eb7c5-7dae-42b5-9b2a-864d053f8c0e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.822363] env[62569]: DEBUG oslo_vmware.api [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Waiting for the task: (returnval){ [ 818.822363] env[62569]: value = "task-1250127" [ 818.822363] env[62569]: _type = "Task" [ 818.822363] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.830855] env[62569]: DEBUG oslo_vmware.api [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Task: {'id': task-1250127, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.919047] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250122, 'name': ReconfigVM_Task, 'duration_secs': 0.606296} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.919340] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Reconfigured VM instance instance-00000047 to attach disk [datastore1] b47cd2d7-0cd2-41af-8ed1-a6dfca323516/b47cd2d7-0cd2-41af-8ed1-a6dfca323516.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 818.919983] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-76af7c6e-bd58-46bf-9307-435dd895df8a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.926680] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 818.926680] env[62569]: value = "task-1250128" [ 818.926680] env[62569]: _type = "Task" [ 818.926680] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.936139] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250128, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.018217] env[62569]: INFO nova.compute.manager [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Took 33.38 seconds to build instance. [ 819.195268] env[62569]: DEBUG nova.network.neutron [req-89f38a72-656c-47fb-a474-3bedfa0c4ea8 req-b903f194-d491-4ade-bd47-05984a8a2261 service nova] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Updated VIF entry in instance network info cache for port cbc4229a-f42d-463c-9a46-6a4d6b3581b0. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 819.195268] env[62569]: DEBUG nova.network.neutron [req-89f38a72-656c-47fb-a474-3bedfa0c4ea8 req-b903f194-d491-4ade-bd47-05984a8a2261 service nova] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Updating instance_info_cache with network_info: [{"id": "cbc4229a-f42d-463c-9a46-6a4d6b3581b0", "address": "fa:16:3e:14:62:67", "network": {"id": "334a2e22-0249-48e3-a2f5-ac37df9771d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1048653043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f7fb3e698324f8e851aefbee74ac669", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f880ac2e-d532-4f54-87bb-998a8d1bca78", "external-id": "nsx-vlan-transportzone-491", "segmentation_id": 491, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcbc4229a-f4", "ovs_interfaceid": "cbc4229a-f42d-463c-9a46-6a4d6b3581b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.296629] env[62569]: DEBUG nova.scheduler.client.report [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 819.312289] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5213d248-db04-100c-0747-6658850b8960, 'name': SearchDatastore_Task, 'duration_secs': 0.020995} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.312659] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.312994] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 819.313449] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.313630] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.313832] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 819.314128] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e0178c02-2fec-46ac-b5bd-650801348fe2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.334014] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 819.334329] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 819.335869] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ba61a2d-5295-4056-b108-b53b3468dfb4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.346928] env[62569]: DEBUG oslo_vmware.api [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Task: {'id': task-1250127, 'name': PowerOffVM_Task, 'duration_secs': 0.19739} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.349015] env[62569]: DEBUG nova.compute.manager [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 819.352280] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 819.352535] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 819.355323] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-640b1485-d1ef-4c1b-9c2b-db3515f2f8f2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.356772] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 819.356772] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52572c27-6860-8a8b-ddd8-81ce7b3df206" [ 819.356772] env[62569]: _type = "Task" [ 819.356772] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.369281] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52572c27-6860-8a8b-ddd8-81ce7b3df206, 'name': SearchDatastore_Task, 'duration_secs': 0.009927} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.371636] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7da1e0c7-1305-42c7-9e73-050f2afd587f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.376997] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 819.376997] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f4e2d5-1297-25ce-208f-bd614023c1ea" [ 819.376997] env[62569]: _type = "Task" [ 819.376997] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.390542] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f4e2d5-1297-25ce-208f-bd614023c1ea, 'name': SearchDatastore_Task, 'duration_secs': 0.00913} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.391090] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.391407] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] cb9d8991-b7ba-4bf7-9d22-1391f4cea768/cb9d8991-b7ba-4bf7-9d22-1391f4cea768.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 819.394023] env[62569]: DEBUG nova.virt.hardware [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 819.394173] env[62569]: DEBUG nova.virt.hardware [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 819.394337] env[62569]: DEBUG nova.virt.hardware [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 819.394584] env[62569]: DEBUG nova.virt.hardware [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 819.394678] env[62569]: DEBUG nova.virt.hardware [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 819.394851] env[62569]: DEBUG nova.virt.hardware [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 819.395143] env[62569]: DEBUG nova.virt.hardware [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 819.395510] env[62569]: DEBUG nova.virt.hardware [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 819.395510] env[62569]: DEBUG nova.virt.hardware [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 819.395596] env[62569]: DEBUG nova.virt.hardware [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 819.395765] env[62569]: DEBUG nova.virt.hardware [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 819.396083] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bd1f2f22-dfcb-48bf-87c0-534355ddd4ad {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.399134] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87992a0f-02f4-474f-86b6-72fd06a1ac61 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.409769] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de5f1a82-e538-4bbc-9b27-b15b84820b48 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.414601] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 819.414601] env[62569]: value = "task-1250130" [ 819.414601] env[62569]: _type = "Task" [ 819.414601] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.436924] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250130, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.442271] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250128, 'name': Rename_Task, 'duration_secs': 0.18125} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.442646] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 819.442928] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a0f9763f-0591-421b-9404-da5657b30f13 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.450568] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 819.450568] env[62569]: value = "task-1250131" [ 819.450568] env[62569]: _type = "Task" [ 819.450568] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.460287] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250131, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.470239] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 819.470620] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 819.471014] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Deleting the datastore file [datastore1] 390bcf25-689d-46ad-bffb-3670c3729397 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 819.471627] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d4f6a47b-82f0-401d-b5a5-a432b461030a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.478775] env[62569]: DEBUG oslo_vmware.api [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Waiting for the task: (returnval){ [ 819.478775] env[62569]: value = "task-1250132" [ 819.478775] env[62569]: _type = "Task" [ 819.478775] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.487109] env[62569]: DEBUG oslo_vmware.api [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Task: {'id': task-1250132, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.521099] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca49b2e9-f061-4043-b2b3-046663e6a921 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Lock "fdeac8b8-a34d-4b4c-8f72-281f84e22ebe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.901s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.697643] env[62569]: DEBUG oslo_concurrency.lockutils [req-89f38a72-656c-47fb-a474-3bedfa0c4ea8 req-b903f194-d491-4ade-bd47-05984a8a2261 service nova] Releasing lock "refresh_cache-cb9d8991-b7ba-4bf7-9d22-1391f4cea768" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.698098] env[62569]: DEBUG nova.compute.manager [req-89f38a72-656c-47fb-a474-3bedfa0c4ea8 req-b903f194-d491-4ade-bd47-05984a8a2261 service nova] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Received event network-vif-deleted-abd9f1dd-a26a-4f38-bee3-041710cbd705 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 819.807782] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.481s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.809755] env[62569]: DEBUG oslo_concurrency.lockutils [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.481s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.810778] env[62569]: INFO nova.compute.claims [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 819.932398] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250130, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.960819] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250131, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.989714] env[62569]: DEBUG oslo_vmware.api [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Task: {'id': task-1250132, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.315675] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Acquiring lock "150accd1-8be0-414f-9d90-c34a34dc4029" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.315948] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Lock "150accd1-8be0-414f-9d90-c34a34dc4029" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.356458] env[62569]: DEBUG nova.compute.manager [req-ec460e09-510d-4055-acf6-b35d7f1de761 req-c72c4f5d-eb59-4059-bbdf-c779b65479b1 service nova] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Received event network-changed-da0de7a0-922e-46b2-842b-a222e31de3b2 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 820.356672] env[62569]: DEBUG nova.compute.manager [req-ec460e09-510d-4055-acf6-b35d7f1de761 req-c72c4f5d-eb59-4059-bbdf-c779b65479b1 service nova] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Refreshing instance network info cache due to event network-changed-da0de7a0-922e-46b2-842b-a222e31de3b2. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 820.356890] env[62569]: DEBUG oslo_concurrency.lockutils [req-ec460e09-510d-4055-acf6-b35d7f1de761 req-c72c4f5d-eb59-4059-bbdf-c779b65479b1 service nova] Acquiring lock "refresh_cache-fdeac8b8-a34d-4b4c-8f72-281f84e22ebe" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.361095] env[62569]: DEBUG oslo_concurrency.lockutils [req-ec460e09-510d-4055-acf6-b35d7f1de761 req-c72c4f5d-eb59-4059-bbdf-c779b65479b1 service nova] Acquired lock "refresh_cache-fdeac8b8-a34d-4b4c-8f72-281f84e22ebe" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.361371] env[62569]: DEBUG nova.network.neutron [req-ec460e09-510d-4055-acf6-b35d7f1de761 req-c72c4f5d-eb59-4059-bbdf-c779b65479b1 service nova] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Refreshing network info cache for port da0de7a0-922e-46b2-842b-a222e31de3b2 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 820.427088] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250130, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.614819} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.427573] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] cb9d8991-b7ba-4bf7-9d22-1391f4cea768/cb9d8991-b7ba-4bf7-9d22-1391f4cea768.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 820.427968] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 820.428367] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-155a8c18-a4b6-4ad6-a710-3bf1e87de685 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.437936] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 820.437936] env[62569]: value = "task-1250133" [ 820.437936] env[62569]: _type = "Task" [ 820.437936] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.448466] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250133, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.461882] env[62569]: DEBUG oslo_vmware.api [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250131, 'name': PowerOnVM_Task, 'duration_secs': 0.773872} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.462222] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 820.462563] env[62569]: INFO nova.compute.manager [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Took 10.36 seconds to spawn the instance on the hypervisor. [ 820.462823] env[62569]: DEBUG nova.compute.manager [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 820.463908] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8ac0222-2f30-4b3e-a410-3692df22a4ba {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.486621] env[62569]: DEBUG nova.compute.manager [req-35faa878-0161-42a1-adfe-bc0350247b50 req-e3e8fb6b-c9e4-4bcc-bb33-997bc9d56719 service nova] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Received event network-vif-plugged-39e71a4f-3b81-4b34-8cc2-2a5352eca4ce {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 820.486937] env[62569]: DEBUG oslo_concurrency.lockutils [req-35faa878-0161-42a1-adfe-bc0350247b50 req-e3e8fb6b-c9e4-4bcc-bb33-997bc9d56719 service nova] Acquiring lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.489212] env[62569]: DEBUG oslo_concurrency.lockutils [req-35faa878-0161-42a1-adfe-bc0350247b50 req-e3e8fb6b-c9e4-4bcc-bb33-997bc9d56719 service nova] Lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.489212] env[62569]: DEBUG oslo_concurrency.lockutils [req-35faa878-0161-42a1-adfe-bc0350247b50 req-e3e8fb6b-c9e4-4bcc-bb33-997bc9d56719 service nova] Lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.489212] env[62569]: DEBUG nova.compute.manager [req-35faa878-0161-42a1-adfe-bc0350247b50 req-e3e8fb6b-c9e4-4bcc-bb33-997bc9d56719 service nova] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] No waiting events found dispatching network-vif-plugged-39e71a4f-3b81-4b34-8cc2-2a5352eca4ce {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 820.489212] env[62569]: WARNING nova.compute.manager [req-35faa878-0161-42a1-adfe-bc0350247b50 req-e3e8fb6b-c9e4-4bcc-bb33-997bc9d56719 service nova] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Received unexpected event network-vif-plugged-39e71a4f-3b81-4b34-8cc2-2a5352eca4ce for instance with vm_state building and task_state spawning. [ 820.495480] env[62569]: DEBUG oslo_vmware.api [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Task: {'id': task-1250132, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.540005} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.495749] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 820.495933] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 820.496169] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 820.499564] env[62569]: INFO nova.compute.manager [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Took 1.69 seconds to destroy the instance on the hypervisor. [ 820.499564] env[62569]: DEBUG oslo.service.loopingcall [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 820.499564] env[62569]: DEBUG nova.compute.manager [-] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 820.499564] env[62569]: DEBUG nova.network.neutron [-] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 820.821448] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Lock "150accd1-8be0-414f-9d90-c34a34dc4029" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.505s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.824088] env[62569]: DEBUG nova.compute.manager [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 820.950912] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250133, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066551} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.953630] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 820.954696] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35deb9b8-c2c5-4e6c-9dac-374f8ae111b3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.978682] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] cb9d8991-b7ba-4bf7-9d22-1391f4cea768/cb9d8991-b7ba-4bf7-9d22-1391f4cea768.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 820.991746] env[62569]: INFO nova.compute.manager [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Took 34.81 seconds to build instance. [ 820.996576] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fccc6f65-20d8-42bb-a5f5-1463ae6676f8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.027486] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 821.027486] env[62569]: value = "task-1250134" [ 821.027486] env[62569]: _type = "Task" [ 821.027486] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.043660] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250134, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.079165] env[62569]: DEBUG nova.network.neutron [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Successfully updated port: 39e71a4f-3b81-4b34-8cc2-2a5352eca4ce {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 821.242513] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94001edd-4fa0-4d6d-a9a3-25430cf744b4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.251277] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e17cbc94-6634-4dc6-83ec-7bc25e6aa598 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.282727] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d78fbb83-23b5-4baa-8e55-0904ae55021c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.291200] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e870c33-3ee7-415b-ab50-db15f333756c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.305768] env[62569]: DEBUG nova.compute.provider_tree [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 821.319345] env[62569]: DEBUG nova.network.neutron [req-ec460e09-510d-4055-acf6-b35d7f1de761 req-c72c4f5d-eb59-4059-bbdf-c779b65479b1 service nova] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Updated VIF entry in instance network info cache for port da0de7a0-922e-46b2-842b-a222e31de3b2. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 821.319744] env[62569]: DEBUG nova.network.neutron [req-ec460e09-510d-4055-acf6-b35d7f1de761 req-c72c4f5d-eb59-4059-bbdf-c779b65479b1 service nova] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Updating instance_info_cache with network_info: [{"id": "da0de7a0-922e-46b2-842b-a222e31de3b2", "address": "fa:16:3e:eb:02:d2", "network": {"id": "ea8253a1-5878-4093-9acf-1426d966475b", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-616956732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.226", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3de498c4aa74ba698deedcacb9b44f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d33839ae-40ca-471b-92e3-eb282b920682", "external-id": "nsx-vlan-transportzone-416", "segmentation_id": 416, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda0de7a0-92", "ovs_interfaceid": "da0de7a0-922e-46b2-842b-a222e31de3b2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.329278] env[62569]: DEBUG nova.compute.utils [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 821.332896] env[62569]: DEBUG nova.compute.manager [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 821.332896] env[62569]: DEBUG nova.network.neutron [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 821.379871] env[62569]: DEBUG nova.policy [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2cea54ce2bb34b9887e04636413c2d72', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a233e091c89d4b8fbaf6e5fb5906b011', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 821.497158] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd36148a-b558-4324-aa21-6544142020a5 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "b47cd2d7-0cd2-41af-8ed1-a6dfca323516" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.316s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.542442] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250134, 'name': ReconfigVM_Task, 'duration_secs': 0.366989} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.542442] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Reconfigured VM instance instance-00000048 to attach disk [datastore1] cb9d8991-b7ba-4bf7-9d22-1391f4cea768/cb9d8991-b7ba-4bf7-9d22-1391f4cea768.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 821.543039] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6780688a-6846-4b17-a800-d6f9c573c806 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.549734] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 821.549734] env[62569]: value = "task-1250135" [ 821.549734] env[62569]: _type = "Task" [ 821.549734] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.564337] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250135, 'name': Rename_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.585869] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.586153] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.586328] env[62569]: DEBUG nova.network.neutron [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 821.615595] env[62569]: DEBUG nova.network.neutron [-] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.807762] env[62569]: DEBUG nova.network.neutron [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Successfully created port: 8e0244f3-9b3e-4fe8-8cfd-666a36f1ef58 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 821.811464] env[62569]: DEBUG nova.scheduler.client.report [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 821.823093] env[62569]: DEBUG oslo_concurrency.lockutils [req-ec460e09-510d-4055-acf6-b35d7f1de761 req-c72c4f5d-eb59-4059-bbdf-c779b65479b1 service nova] Releasing lock "refresh_cache-fdeac8b8-a34d-4b4c-8f72-281f84e22ebe" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.834527] env[62569]: DEBUG nova.compute.manager [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 822.060283] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250135, 'name': Rename_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.119860] env[62569]: DEBUG nova.network.neutron [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 822.122108] env[62569]: INFO nova.compute.manager [-] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Took 1.63 seconds to deallocate network for instance. [ 822.272031] env[62569]: DEBUG nova.network.neutron [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Updating instance_info_cache with network_info: [{"id": "39e71a4f-3b81-4b34-8cc2-2a5352eca4ce", "address": "fa:16:3e:4b:e2:a8", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap39e71a4f-3b", "ovs_interfaceid": "39e71a4f-3b81-4b34-8cc2-2a5352eca4ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.323857] env[62569]: DEBUG oslo_concurrency.lockutils [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.515s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.324417] env[62569]: DEBUG nova.compute.manager [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 822.327056] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.175s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.327282] env[62569]: DEBUG nova.objects.instance [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lazy-loading 'resources' on Instance uuid 37a4c60a-e660-4c91-bcfe-72638a4397b0 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 822.395750] env[62569]: DEBUG nova.compute.manager [req-f1be3287-944c-4468-999a-077343480e1e req-7dc3e24c-af08-4f8b-990e-4417298c283f service nova] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Received event network-vif-deleted-a466b80c-87f5-4bfd-ad2a-fc3610a634ed {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 822.396452] env[62569]: DEBUG nova.compute.manager [req-f1be3287-944c-4468-999a-077343480e1e req-7dc3e24c-af08-4f8b-990e-4417298c283f service nova] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Received event network-changed-39e71a4f-3b81-4b34-8cc2-2a5352eca4ce {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 822.396452] env[62569]: DEBUG nova.compute.manager [req-f1be3287-944c-4468-999a-077343480e1e req-7dc3e24c-af08-4f8b-990e-4417298c283f service nova] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Refreshing instance network info cache due to event network-changed-39e71a4f-3b81-4b34-8cc2-2a5352eca4ce. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 822.396602] env[62569]: DEBUG oslo_concurrency.lockutils [req-f1be3287-944c-4468-999a-077343480e1e req-7dc3e24c-af08-4f8b-990e-4417298c283f service nova] Acquiring lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.560588] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250135, 'name': Rename_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.635281] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.775044] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.775411] env[62569]: DEBUG nova.compute.manager [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Instance network_info: |[{"id": "39e71a4f-3b81-4b34-8cc2-2a5352eca4ce", "address": "fa:16:3e:4b:e2:a8", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap39e71a4f-3b", "ovs_interfaceid": "39e71a4f-3b81-4b34-8cc2-2a5352eca4ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 822.775742] env[62569]: DEBUG oslo_concurrency.lockutils [req-f1be3287-944c-4468-999a-077343480e1e req-7dc3e24c-af08-4f8b-990e-4417298c283f service nova] Acquired lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.775934] env[62569]: DEBUG nova.network.neutron [req-f1be3287-944c-4468-999a-077343480e1e req-7dc3e24c-af08-4f8b-990e-4417298c283f service nova] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Refreshing network info cache for port 39e71a4f-3b81-4b34-8cc2-2a5352eca4ce {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 822.777390] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4b:e2:a8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73915082-a1b0-460b-b24d-97588fc9cb29', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '39e71a4f-3b81-4b34-8cc2-2a5352eca4ce', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 822.785346] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Creating folder: Project (7f988acdb9cd4ca28ca8916c2ede2db5). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 822.786495] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-200d535d-2bc3-41ef-8e55-f10b79cf8922 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.798016] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Created folder: Project (7f988acdb9cd4ca28ca8916c2ede2db5) in parent group-v269330. [ 822.798230] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Creating folder: Instances. Parent ref: group-v269421. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 822.798474] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-251756aa-2544-40e6-ba38-1bf1511eade8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.808366] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Created folder: Instances in parent group-v269421. [ 822.808636] env[62569]: DEBUG oslo.service.loopingcall [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.808843] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 822.809071] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9905db80-0879-4fb1-af5b-8f887ebc2645 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.832547] env[62569]: DEBUG nova.compute.utils [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 822.833950] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 822.833950] env[62569]: value = "task-1250138" [ 822.833950] env[62569]: _type = "Task" [ 822.833950] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.834452] env[62569]: DEBUG nova.compute.manager [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 822.834621] env[62569]: DEBUG nova.network.neutron [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 822.847062] env[62569]: DEBUG nova.compute.manager [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 822.849066] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250138, 'name': CreateVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.890223] env[62569]: DEBUG nova.virt.hardware [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 822.890223] env[62569]: DEBUG nova.virt.hardware [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 822.890223] env[62569]: DEBUG nova.virt.hardware [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 822.890486] env[62569]: DEBUG nova.virt.hardware [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 822.890486] env[62569]: DEBUG nova.virt.hardware [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 822.890653] env[62569]: DEBUG nova.virt.hardware [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 822.890930] env[62569]: DEBUG nova.virt.hardware [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 822.891125] env[62569]: DEBUG nova.virt.hardware [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 822.891300] env[62569]: DEBUG nova.virt.hardware [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 822.891549] env[62569]: DEBUG nova.virt.hardware [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 822.891814] env[62569]: DEBUG nova.virt.hardware [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 822.893071] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca4230d-7f6a-42e9-a732-9ea8445d3308 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.906252] env[62569]: DEBUG nova.policy [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e5307b1d2b69421ab908a0d655a593fe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'de8605118d744a93bce54e897aa849b5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 822.911259] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ee7dbdb-cc6e-4c84-a8be-f729c4dce523 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.061891] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250135, 'name': Rename_Task, 'duration_secs': 1.402144} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.062262] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 823.062550] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-db511f71-9bc1-4c81-811f-6a07529dd551 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.070537] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 823.070537] env[62569]: value = "task-1250139" [ 823.070537] env[62569]: _type = "Task" [ 823.070537] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.078869] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250139, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.229390] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1ee6f82-ac15-4db3-a1ea-9c10335ec581 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.235444] env[62569]: DEBUG nova.network.neutron [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Successfully created port: 174411d8-9538-4849-b0d1-fdb27fe71a7b {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 823.241030] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61861819-d7f6-4d6d-bd55-5f93aa3963ca {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.273623] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b4eb4e4-4dc4-42b6-8b54-ea8b4a80e7c1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.281296] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6483141f-22d6-4f86-bbf3-3ee16ad7388d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.297093] env[62569]: DEBUG nova.compute.provider_tree [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 823.337966] env[62569]: DEBUG nova.compute.manager [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 823.351961] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250138, 'name': CreateVM_Task, 'duration_secs': 0.448035} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.352092] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 823.354610] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.354610] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.354610] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 823.354610] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fbb3f903-12b4-4997-bf95-22b074390b91 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.360021] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 823.360021] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b249fe-0dab-ac09-4da5-52164c3f8596" [ 823.360021] env[62569]: _type = "Task" [ 823.360021] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.367925] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b249fe-0dab-ac09-4da5-52164c3f8596, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.428385] env[62569]: DEBUG nova.network.neutron [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Successfully updated port: 8e0244f3-9b3e-4fe8-8cfd-666a36f1ef58 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 823.435819] env[62569]: DEBUG nova.compute.manager [req-e7e28937-5cd7-48d9-9be8-c15607cee8da req-d14fc831-d86d-4072-9e6d-5afb27ab706b service nova] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Received event network-vif-plugged-8e0244f3-9b3e-4fe8-8cfd-666a36f1ef58 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 823.436118] env[62569]: DEBUG oslo_concurrency.lockutils [req-e7e28937-5cd7-48d9-9be8-c15607cee8da req-d14fc831-d86d-4072-9e6d-5afb27ab706b service nova] Acquiring lock "6e83fd32-215a-42e4-99e7-72bdce25555e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.436353] env[62569]: DEBUG oslo_concurrency.lockutils [req-e7e28937-5cd7-48d9-9be8-c15607cee8da req-d14fc831-d86d-4072-9e6d-5afb27ab706b service nova] Lock "6e83fd32-215a-42e4-99e7-72bdce25555e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.436533] env[62569]: DEBUG oslo_concurrency.lockutils [req-e7e28937-5cd7-48d9-9be8-c15607cee8da req-d14fc831-d86d-4072-9e6d-5afb27ab706b service nova] Lock "6e83fd32-215a-42e4-99e7-72bdce25555e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.436790] env[62569]: DEBUG nova.compute.manager [req-e7e28937-5cd7-48d9-9be8-c15607cee8da req-d14fc831-d86d-4072-9e6d-5afb27ab706b service nova] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] No waiting events found dispatching network-vif-plugged-8e0244f3-9b3e-4fe8-8cfd-666a36f1ef58 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 823.436980] env[62569]: WARNING nova.compute.manager [req-e7e28937-5cd7-48d9-9be8-c15607cee8da req-d14fc831-d86d-4072-9e6d-5afb27ab706b service nova] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Received unexpected event network-vif-plugged-8e0244f3-9b3e-4fe8-8cfd-666a36f1ef58 for instance with vm_state building and task_state spawning. [ 823.585166] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250139, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.632299] env[62569]: DEBUG nova.network.neutron [req-f1be3287-944c-4468-999a-077343480e1e req-7dc3e24c-af08-4f8b-990e-4417298c283f service nova] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Updated VIF entry in instance network info cache for port 39e71a4f-3b81-4b34-8cc2-2a5352eca4ce. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 823.632692] env[62569]: DEBUG nova.network.neutron [req-f1be3287-944c-4468-999a-077343480e1e req-7dc3e24c-af08-4f8b-990e-4417298c283f service nova] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Updating instance_info_cache with network_info: [{"id": "39e71a4f-3b81-4b34-8cc2-2a5352eca4ce", "address": "fa:16:3e:4b:e2:a8", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap39e71a4f-3b", "ovs_interfaceid": "39e71a4f-3b81-4b34-8cc2-2a5352eca4ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.800903] env[62569]: DEBUG nova.scheduler.client.report [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 823.867917] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b249fe-0dab-ac09-4da5-52164c3f8596, 'name': SearchDatastore_Task, 'duration_secs': 0.010471} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.868272] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.868507] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 823.868744] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.868896] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.869094] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 823.869365] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e7a918d2-c298-486c-8533-790862aff874 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.877459] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 823.877646] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 823.878963] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ba428ac-4eec-4625-9a43-598b2e7e1c88 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.884520] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 823.884520] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5229a2c3-a071-0d65-498c-afa71e65e9ea" [ 823.884520] env[62569]: _type = "Task" [ 823.884520] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.892009] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5229a2c3-a071-0d65-498c-afa71e65e9ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.930856] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Acquiring lock "refresh_cache-6e83fd32-215a-42e4-99e7-72bdce25555e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.931026] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Acquired lock "refresh_cache-6e83fd32-215a-42e4-99e7-72bdce25555e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.931193] env[62569]: DEBUG nova.network.neutron [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 824.083106] env[62569]: DEBUG oslo_vmware.api [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250139, 'name': PowerOnVM_Task, 'duration_secs': 0.956166} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.083440] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 824.083686] env[62569]: INFO nova.compute.manager [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Took 9.35 seconds to spawn the instance on the hypervisor. [ 824.083907] env[62569]: DEBUG nova.compute.manager [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 824.084707] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f8a222d-914f-41e9-99c8-463020631800 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.137058] env[62569]: DEBUG oslo_concurrency.lockutils [req-f1be3287-944c-4468-999a-077343480e1e req-7dc3e24c-af08-4f8b-990e-4417298c283f service nova] Releasing lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.137058] env[62569]: DEBUG nova.compute.manager [req-f1be3287-944c-4468-999a-077343480e1e req-7dc3e24c-af08-4f8b-990e-4417298c283f service nova] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Received event network-vif-deleted-ac9a07c1-60a1-4fd4-83c7-54ffc30240d8 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 824.137058] env[62569]: DEBUG nova.compute.manager [req-f1be3287-944c-4468-999a-077343480e1e req-7dc3e24c-af08-4f8b-990e-4417298c283f service nova] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Received event network-changed-90b3aaef-2e72-486f-be8d-a4202a6cb0c6 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 824.137258] env[62569]: DEBUG nova.compute.manager [req-f1be3287-944c-4468-999a-077343480e1e req-7dc3e24c-af08-4f8b-990e-4417298c283f service nova] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Refreshing instance network info cache due to event network-changed-90b3aaef-2e72-486f-be8d-a4202a6cb0c6. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 824.137443] env[62569]: DEBUG oslo_concurrency.lockutils [req-f1be3287-944c-4468-999a-077343480e1e req-7dc3e24c-af08-4f8b-990e-4417298c283f service nova] Acquiring lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.137586] env[62569]: DEBUG oslo_concurrency.lockutils [req-f1be3287-944c-4468-999a-077343480e1e req-7dc3e24c-af08-4f8b-990e-4417298c283f service nova] Acquired lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.137750] env[62569]: DEBUG nova.network.neutron [req-f1be3287-944c-4468-999a-077343480e1e req-7dc3e24c-af08-4f8b-990e-4417298c283f service nova] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Refreshing network info cache for port 90b3aaef-2e72-486f-be8d-a4202a6cb0c6 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 824.305638] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.978s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.308847] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.648s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.309136] env[62569]: DEBUG nova.objects.instance [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lazy-loading 'resources' on Instance uuid 06067572-a9fd-43a4-91cd-383a62f53885 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 824.328829] env[62569]: INFO nova.scheduler.client.report [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Deleted allocations for instance 37a4c60a-e660-4c91-bcfe-72638a4397b0 [ 824.349809] env[62569]: DEBUG nova.compute.manager [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 824.374337] env[62569]: DEBUG nova.virt.hardware [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 824.374564] env[62569]: DEBUG nova.virt.hardware [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 824.374727] env[62569]: DEBUG nova.virt.hardware [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 824.374917] env[62569]: DEBUG nova.virt.hardware [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 824.375077] env[62569]: DEBUG nova.virt.hardware [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 824.375231] env[62569]: DEBUG nova.virt.hardware [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 824.375446] env[62569]: DEBUG nova.virt.hardware [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 824.375606] env[62569]: DEBUG nova.virt.hardware [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 824.375815] env[62569]: DEBUG nova.virt.hardware [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 824.375990] env[62569]: DEBUG nova.virt.hardware [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 824.376181] env[62569]: DEBUG nova.virt.hardware [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 824.377064] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f04ce50-1baf-45cc-93ab-2acab1e71000 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.385605] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0079c85-f06f-4cfa-9269-26ff695ce124 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.405873] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5229a2c3-a071-0d65-498c-afa71e65e9ea, 'name': SearchDatastore_Task, 'duration_secs': 0.012286} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.406528] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f7f30b4-6ddf-4bbe-851a-a61d7c09ead7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.411419] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 824.411419] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5237f7fe-8a71-60b1-323f-508ea92acf2c" [ 824.411419] env[62569]: _type = "Task" [ 824.411419] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.418506] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5237f7fe-8a71-60b1-323f-508ea92acf2c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.462445] env[62569]: DEBUG nova.network.neutron [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.594426] env[62569]: DEBUG nova.network.neutron [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Updating instance_info_cache with network_info: [{"id": "8e0244f3-9b3e-4fe8-8cfd-666a36f1ef58", "address": "fa:16:3e:27:83:88", "network": {"id": "4d7a88ec-5e0d-490d-9852-4e69b9ce920d", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-933452324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a233e091c89d4b8fbaf6e5fb5906b011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e0244f3-9b", "ovs_interfaceid": "8e0244f3-9b3e-4fe8-8cfd-666a36f1ef58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.604061] env[62569]: INFO nova.compute.manager [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Took 34.82 seconds to build instance. [ 824.840410] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1dcbe57a-1a28-4f86-a3af-eb52809c97e2 tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "37a4c60a-e660-4c91-bcfe-72638a4397b0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.436s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.921610] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5237f7fe-8a71-60b1-323f-508ea92acf2c, 'name': SearchDatastore_Task, 'duration_secs': 0.009304} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.924479] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.924840] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] e4eadcdf-a04a-4255-ba1c-fe20156c655f/e4eadcdf-a04a-4255-ba1c-fe20156c655f.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 824.925321] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ba007969-a732-4708-8b79-9375af6594bd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.932093] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 824.932093] env[62569]: value = "task-1250140" [ 824.932093] env[62569]: _type = "Task" [ 824.932093] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.940507] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250140, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.986239] env[62569]: DEBUG nova.compute.manager [req-e349d6e0-6c76-4392-b99c-28176b2ddea3 req-20362478-2622-4760-8ab5-fcea072aaf45 service nova] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Received event network-vif-plugged-174411d8-9538-4849-b0d1-fdb27fe71a7b {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 824.987095] env[62569]: DEBUG oslo_concurrency.lockutils [req-e349d6e0-6c76-4392-b99c-28176b2ddea3 req-20362478-2622-4760-8ab5-fcea072aaf45 service nova] Acquiring lock "c9568fcb-745f-4c04-8487-a38a617aab7c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.987721] env[62569]: DEBUG oslo_concurrency.lockutils [req-e349d6e0-6c76-4392-b99c-28176b2ddea3 req-20362478-2622-4760-8ab5-fcea072aaf45 service nova] Lock "c9568fcb-745f-4c04-8487-a38a617aab7c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.988149] env[62569]: DEBUG oslo_concurrency.lockutils [req-e349d6e0-6c76-4392-b99c-28176b2ddea3 req-20362478-2622-4760-8ab5-fcea072aaf45 service nova] Lock "c9568fcb-745f-4c04-8487-a38a617aab7c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.988761] env[62569]: DEBUG nova.compute.manager [req-e349d6e0-6c76-4392-b99c-28176b2ddea3 req-20362478-2622-4760-8ab5-fcea072aaf45 service nova] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] No waiting events found dispatching network-vif-plugged-174411d8-9538-4849-b0d1-fdb27fe71a7b {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 824.989152] env[62569]: WARNING nova.compute.manager [req-e349d6e0-6c76-4392-b99c-28176b2ddea3 req-20362478-2622-4760-8ab5-fcea072aaf45 service nova] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Received unexpected event network-vif-plugged-174411d8-9538-4849-b0d1-fdb27fe71a7b for instance with vm_state building and task_state spawning. [ 825.007884] env[62569]: DEBUG nova.network.neutron [req-f1be3287-944c-4468-999a-077343480e1e req-7dc3e24c-af08-4f8b-990e-4417298c283f service nova] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Updated VIF entry in instance network info cache for port 90b3aaef-2e72-486f-be8d-a4202a6cb0c6. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 825.008545] env[62569]: DEBUG nova.network.neutron [req-f1be3287-944c-4468-999a-077343480e1e req-7dc3e24c-af08-4f8b-990e-4417298c283f service nova] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Updating instance_info_cache with network_info: [{"id": "90b3aaef-2e72-486f-be8d-a4202a6cb0c6", "address": "fa:16:3e:0c:8f:dd", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90b3aaef-2e", "ovs_interfaceid": "90b3aaef-2e72-486f-be8d-a4202a6cb0c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.059057] env[62569]: DEBUG nova.network.neutron [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Successfully updated port: 174411d8-9538-4849-b0d1-fdb27fe71a7b {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 825.101405] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Releasing lock "refresh_cache-6e83fd32-215a-42e4-99e7-72bdce25555e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.101405] env[62569]: DEBUG nova.compute.manager [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Instance network_info: |[{"id": "8e0244f3-9b3e-4fe8-8cfd-666a36f1ef58", "address": "fa:16:3e:27:83:88", "network": {"id": "4d7a88ec-5e0d-490d-9852-4e69b9ce920d", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-933452324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a233e091c89d4b8fbaf6e5fb5906b011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e0244f3-9b", "ovs_interfaceid": "8e0244f3-9b3e-4fe8-8cfd-666a36f1ef58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 825.104533] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:27:83:88', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '62f28d75-4e6a-4ae5-b8b3-d0652ea26d08', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8e0244f3-9b3e-4fe8-8cfd-666a36f1ef58', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 825.112252] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Creating folder: Project (a233e091c89d4b8fbaf6e5fb5906b011). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 825.113141] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6ecf5200-6333-4340-8cfa-76046a2dd4b9 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cb9d8991-b7ba-4bf7-9d22-1391f4cea768" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.205s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.113746] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6ad614df-f29f-4296-8077-ca1659952407 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.127781] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Created folder: Project (a233e091c89d4b8fbaf6e5fb5906b011) in parent group-v269330. [ 825.128028] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Creating folder: Instances. Parent ref: group-v269424. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 825.128326] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3ee5d97e-7e91-4bb6-bc9f-b16e54be453c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.137926] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Created folder: Instances in parent group-v269424. [ 825.138262] env[62569]: DEBUG oslo.service.loopingcall [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.141295] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 825.141814] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-15ef1569-9fd4-46b0-badb-246b1cdcbf24 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.166629] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 825.166629] env[62569]: value = "task-1250143" [ 825.166629] env[62569]: _type = "Task" [ 825.166629] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.177209] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250143, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.188185] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30466fd4-5b46-4197-b1cb-e9fb651bab71 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.197082] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43f221a3-2bb4-4054-9d96-57c8c6421cc1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.231859] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e343034-6257-40df-b199-5e59b640931f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.240225] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2dcf49e-729b-4cf2-9b57-632085fd9029 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.255201] env[62569]: DEBUG nova.compute.provider_tree [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 825.443451] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250140, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457136} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.444118] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] e4eadcdf-a04a-4255-ba1c-fe20156c655f/e4eadcdf-a04a-4255-ba1c-fe20156c655f.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 825.445060] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 825.445060] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5cf1e09e-6484-4879-ac3d-73c27be4fe97 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.451416] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 825.451416] env[62569]: value = "task-1250144" [ 825.451416] env[62569]: _type = "Task" [ 825.451416] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.451830] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 825.455056] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 825.460493] env[62569]: DEBUG nova.compute.manager [req-af4dbad6-f9de-4920-87c6-7e0d0065d417 req-e2c0c886-b61d-4117-9e74-183ab7f250e1 service nova] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Received event network-changed-8e0244f3-9b3e-4fe8-8cfd-666a36f1ef58 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 825.460493] env[62569]: DEBUG nova.compute.manager [req-af4dbad6-f9de-4920-87c6-7e0d0065d417 req-e2c0c886-b61d-4117-9e74-183ab7f250e1 service nova] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Refreshing instance network info cache due to event network-changed-8e0244f3-9b3e-4fe8-8cfd-666a36f1ef58. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 825.460815] env[62569]: DEBUG oslo_concurrency.lockutils [req-af4dbad6-f9de-4920-87c6-7e0d0065d417 req-e2c0c886-b61d-4117-9e74-183ab7f250e1 service nova] Acquiring lock "refresh_cache-6e83fd32-215a-42e4-99e7-72bdce25555e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.461155] env[62569]: DEBUG oslo_concurrency.lockutils [req-af4dbad6-f9de-4920-87c6-7e0d0065d417 req-e2c0c886-b61d-4117-9e74-183ab7f250e1 service nova] Acquired lock "refresh_cache-6e83fd32-215a-42e4-99e7-72bdce25555e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.461637] env[62569]: DEBUG nova.network.neutron [req-af4dbad6-f9de-4920-87c6-7e0d0065d417 req-e2c0c886-b61d-4117-9e74-183ab7f250e1 service nova] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Refreshing network info cache for port 8e0244f3-9b3e-4fe8-8cfd-666a36f1ef58 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 825.466668] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250144, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.511876] env[62569]: DEBUG oslo_concurrency.lockutils [req-f1be3287-944c-4468-999a-077343480e1e req-7dc3e24c-af08-4f8b-990e-4417298c283f service nova] Releasing lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.561973] env[62569]: DEBUG oslo_concurrency.lockutils [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "refresh_cache-c9568fcb-745f-4c04-8487-a38a617aab7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.562221] env[62569]: DEBUG oslo_concurrency.lockutils [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquired lock "refresh_cache-c9568fcb-745f-4c04-8487-a38a617aab7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.562289] env[62569]: DEBUG nova.network.neutron [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 825.676603] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250143, 'name': CreateVM_Task, 'duration_secs': 0.361196} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.677515] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 825.677515] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.677646] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.678242] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 825.678326] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dcececf8-ddd1-43fd-b67d-bda0f25f229d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.682911] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Waiting for the task: (returnval){ [ 825.682911] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d9e8e4-21a5-5884-1b0e-bcf75b3d909d" [ 825.682911] env[62569]: _type = "Task" [ 825.682911] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.692264] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d9e8e4-21a5-5884-1b0e-bcf75b3d909d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.758514] env[62569]: DEBUG nova.scheduler.client.report [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 825.967073] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250144, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06174} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.967073] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 825.967073] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 825.967073] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Starting heal instance info cache {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 825.967073] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Rebuilding the list of instances to heal {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10312}} [ 825.967749] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26f4215f-0bbb-4c94-9d6d-c599640bd477 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.994161] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] e4eadcdf-a04a-4255-ba1c-fe20156c655f/e4eadcdf-a04a-4255-ba1c-fe20156c655f.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 825.995105] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f25d06db-00b7-4239-bb8c-ee2d285d3951 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.019440] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 826.019440] env[62569]: value = "task-1250145" [ 826.019440] env[62569]: _type = "Task" [ 826.019440] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.029964] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250145, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.114089] env[62569]: DEBUG nova.network.neutron [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 826.192694] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d9e8e4-21a5-5884-1b0e-bcf75b3d909d, 'name': SearchDatastore_Task, 'duration_secs': 0.087114} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.193045] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.193307] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 826.193549] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.193797] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.193944] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 826.194228] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-616573f1-c35d-46c4-b43a-cb19350fa133 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.204451] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 826.204638] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 826.205448] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-620fa899-5ecf-4dd4-9bf8-e0684e945fff {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.210397] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Waiting for the task: (returnval){ [ 826.210397] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5223ad12-2d4b-8763-f984-11f1d7f354d6" [ 826.210397] env[62569]: _type = "Task" [ 826.210397] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.220096] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5223ad12-2d4b-8763-f984-11f1d7f354d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.263368] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.955s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.265661] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.877s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.265935] env[62569]: DEBUG nova.objects.instance [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lazy-loading 'resources' on Instance uuid 94392d48-223d-4205-9627-25488a468769 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 826.299794] env[62569]: INFO nova.scheduler.client.report [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Deleted allocations for instance 06067572-a9fd-43a4-91cd-383a62f53885 [ 826.301485] env[62569]: DEBUG nova.network.neutron [req-af4dbad6-f9de-4920-87c6-7e0d0065d417 req-e2c0c886-b61d-4117-9e74-183ab7f250e1 service nova] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Updated VIF entry in instance network info cache for port 8e0244f3-9b3e-4fe8-8cfd-666a36f1ef58. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 826.301797] env[62569]: DEBUG nova.network.neutron [req-af4dbad6-f9de-4920-87c6-7e0d0065d417 req-e2c0c886-b61d-4117-9e74-183ab7f250e1 service nova] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Updating instance_info_cache with network_info: [{"id": "8e0244f3-9b3e-4fe8-8cfd-666a36f1ef58", "address": "fa:16:3e:27:83:88", "network": {"id": "4d7a88ec-5e0d-490d-9852-4e69b9ce920d", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-933452324-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a233e091c89d4b8fbaf6e5fb5906b011", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e0244f3-9b", "ovs_interfaceid": "8e0244f3-9b3e-4fe8-8cfd-666a36f1ef58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.322222] env[62569]: DEBUG nova.network.neutron [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Updating instance_info_cache with network_info: [{"id": "174411d8-9538-4849-b0d1-fdb27fe71a7b", "address": "fa:16:3e:06:4a:35", "network": {"id": "406dd658-8168-4e73-9b97-861929bfaa2e", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-693938985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de8605118d744a93bce54e897aa849b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap174411d8-95", "ovs_interfaceid": "174411d8-9538-4849-b0d1-fdb27fe71a7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.476703] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Skipping network cache update for instance because it is Building. {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 826.477095] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Skipping network cache update for instance because it is Building. {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 826.477095] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Skipping network cache update for instance because it is Building. {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 826.524832] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "refresh_cache-a0b406c3-9466-41bd-9de1-e675cab2ceef" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.524987] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquired lock "refresh_cache-a0b406c3-9466-41bd-9de1-e675cab2ceef" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.525157] env[62569]: DEBUG nova.network.neutron [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Forcefully refreshing network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 826.525316] env[62569]: DEBUG nova.objects.instance [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lazy-loading 'info_cache' on Instance uuid a0b406c3-9466-41bd-9de1-e675cab2ceef {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 826.529745] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250145, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.721058] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5223ad12-2d4b-8763-f984-11f1d7f354d6, 'name': SearchDatastore_Task, 'duration_secs': 0.027275} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.721874] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-490bb21f-1e0c-4907-bde9-f3204048f331 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.727462] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Waiting for the task: (returnval){ [ 826.727462] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5251abf4-2a01-ab94-c44d-479935d03fe8" [ 826.727462] env[62569]: _type = "Task" [ 826.727462] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.738087] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5251abf4-2a01-ab94-c44d-479935d03fe8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.808756] env[62569]: DEBUG oslo_concurrency.lockutils [req-af4dbad6-f9de-4920-87c6-7e0d0065d417 req-e2c0c886-b61d-4117-9e74-183ab7f250e1 service nova] Releasing lock "refresh_cache-6e83fd32-215a-42e4-99e7-72bdce25555e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.813134] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e550af1c-7196-445b-b28e-069f2fc8abde tempest-MultipleCreateTestJSON-1489131804 tempest-MultipleCreateTestJSON-1489131804-project-member] Lock "06067572-a9fd-43a4-91cd-383a62f53885" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.614s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.824200] env[62569]: DEBUG oslo_concurrency.lockutils [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Releasing lock "refresh_cache-c9568fcb-745f-4c04-8487-a38a617aab7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.824420] env[62569]: DEBUG nova.compute.manager [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Instance network_info: |[{"id": "174411d8-9538-4849-b0d1-fdb27fe71a7b", "address": "fa:16:3e:06:4a:35", "network": {"id": "406dd658-8168-4e73-9b97-861929bfaa2e", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-693938985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de8605118d744a93bce54e897aa849b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap174411d8-95", "ovs_interfaceid": "174411d8-9538-4849-b0d1-fdb27fe71a7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 826.827563] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:06:4a:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '257e5ea7-8b80-4301-9900-a754f1fe2031', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '174411d8-9538-4849-b0d1-fdb27fe71a7b', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 826.835750] env[62569]: DEBUG oslo.service.loopingcall [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 826.836899] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 826.837175] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4261f3c2-8bfa-496e-a218-1cd5e57b7486 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.860644] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 826.860644] env[62569]: value = "task-1250146" [ 826.860644] env[62569]: _type = "Task" [ 826.860644] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.871689] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250146, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.029869] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250145, 'name': ReconfigVM_Task, 'duration_secs': 0.860787} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.033017] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Reconfigured VM instance instance-00000049 to attach disk [datastore1] e4eadcdf-a04a-4255-ba1c-fe20156c655f/e4eadcdf-a04a-4255-ba1c-fe20156c655f.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 827.034601] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ceb32e9c-b31d-49e4-a1bd-6d1d9fc7384d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.040841] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 827.040841] env[62569]: value = "task-1250147" [ 827.040841] env[62569]: _type = "Task" [ 827.040841] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.048515] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250147, 'name': Rename_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.079523] env[62569]: DEBUG nova.compute.manager [req-17bccfbd-0bfb-40d5-bda8-89177b8f36c4 req-2f39d5a0-91e3-45ad-a86d-e3c498db62a7 service nova] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Received event network-changed-174411d8-9538-4849-b0d1-fdb27fe71a7b {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 827.079922] env[62569]: DEBUG nova.compute.manager [req-17bccfbd-0bfb-40d5-bda8-89177b8f36c4 req-2f39d5a0-91e3-45ad-a86d-e3c498db62a7 service nova] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Refreshing instance network info cache due to event network-changed-174411d8-9538-4849-b0d1-fdb27fe71a7b. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 827.080485] env[62569]: DEBUG oslo_concurrency.lockutils [req-17bccfbd-0bfb-40d5-bda8-89177b8f36c4 req-2f39d5a0-91e3-45ad-a86d-e3c498db62a7 service nova] Acquiring lock "refresh_cache-c9568fcb-745f-4c04-8487-a38a617aab7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.080788] env[62569]: DEBUG oslo_concurrency.lockutils [req-17bccfbd-0bfb-40d5-bda8-89177b8f36c4 req-2f39d5a0-91e3-45ad-a86d-e3c498db62a7 service nova] Acquired lock "refresh_cache-c9568fcb-745f-4c04-8487-a38a617aab7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.081246] env[62569]: DEBUG nova.network.neutron [req-17bccfbd-0bfb-40d5-bda8-89177b8f36c4 req-2f39d5a0-91e3-45ad-a86d-e3c498db62a7 service nova] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Refreshing network info cache for port 174411d8-9538-4849-b0d1-fdb27fe71a7b {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 827.085094] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6e5ffdf-5ddb-4ada-ae0b-1b054716562e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.096916] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-063f9d9a-aada-4b46-bdd4-d2ac157e542e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.136408] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db10267e-527f-4228-85fb-6add0772da38 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.144450] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6ec278b-e049-4f9b-87dd-bc2e0e1f9f44 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.163103] env[62569]: DEBUG nova.compute.provider_tree [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 827.242733] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5251abf4-2a01-ab94-c44d-479935d03fe8, 'name': SearchDatastore_Task, 'duration_secs': 0.01249} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.243150] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.243555] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 6e83fd32-215a-42e4-99e7-72bdce25555e/6e83fd32-215a-42e4-99e7-72bdce25555e.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 827.243959] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c7a62939-3248-4ce4-a591-f448b0e6e048 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.251805] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Waiting for the task: (returnval){ [ 827.251805] env[62569]: value = "task-1250148" [ 827.251805] env[62569]: _type = "Task" [ 827.251805] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.263250] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Task: {'id': task-1250148, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.371323] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250146, 'name': CreateVM_Task, 'duration_secs': 0.428441} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.371555] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 827.372321] env[62569]: DEBUG oslo_concurrency.lockutils [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.372729] env[62569]: DEBUG oslo_concurrency.lockutils [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.373136] env[62569]: DEBUG oslo_concurrency.lockutils [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 827.373440] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0bd12d74-8fd5-4c55-b726-e0d4b071277b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.378182] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 827.378182] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52fa94dd-c454-7a9c-88a6-a3e590dba881" [ 827.378182] env[62569]: _type = "Task" [ 827.378182] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.386869] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52fa94dd-c454-7a9c-88a6-a3e590dba881, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.550499] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250147, 'name': Rename_Task, 'duration_secs': 0.184724} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.550901] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 827.553024] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a15f202-8afa-48df-951e-0e24b21bdb0e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.557671] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 827.557671] env[62569]: value = "task-1250149" [ 827.557671] env[62569]: _type = "Task" [ 827.557671] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.565533] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250149, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.666317] env[62569]: DEBUG nova.scheduler.client.report [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 827.763092] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Task: {'id': task-1250148, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.889270] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52fa94dd-c454-7a9c-88a6-a3e590dba881, 'name': SearchDatastore_Task, 'duration_secs': 0.008535} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.891683] env[62569]: DEBUG oslo_concurrency.lockutils [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.891935] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 827.892194] env[62569]: DEBUG oslo_concurrency.lockutils [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.892350] env[62569]: DEBUG oslo_concurrency.lockutils [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.892533] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 827.892819] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-940cca55-13b6-48fb-8cc8-b7267ac96479 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.901665] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 827.901862] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 827.902622] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1cb88f0-1a88-4809-9779-3d0b4a42de99 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.909537] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 827.909537] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f9840f-e26f-035d-bbb7-09742e45ceda" [ 827.909537] env[62569]: _type = "Task" [ 827.909537] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.917410] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f9840f-e26f-035d-bbb7-09742e45ceda, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.992074] env[62569]: DEBUG nova.network.neutron [req-17bccfbd-0bfb-40d5-bda8-89177b8f36c4 req-2f39d5a0-91e3-45ad-a86d-e3c498db62a7 service nova] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Updated VIF entry in instance network info cache for port 174411d8-9538-4849-b0d1-fdb27fe71a7b. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 827.992074] env[62569]: DEBUG nova.network.neutron [req-17bccfbd-0bfb-40d5-bda8-89177b8f36c4 req-2f39d5a0-91e3-45ad-a86d-e3c498db62a7 service nova] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Updating instance_info_cache with network_info: [{"id": "174411d8-9538-4849-b0d1-fdb27fe71a7b", "address": "fa:16:3e:06:4a:35", "network": {"id": "406dd658-8168-4e73-9b97-861929bfaa2e", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-693938985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de8605118d744a93bce54e897aa849b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap174411d8-95", "ovs_interfaceid": "174411d8-9538-4849-b0d1-fdb27fe71a7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.066757] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250149, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.172448] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.906s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.176548] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.098s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.178875] env[62569]: INFO nova.compute.claims [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 828.208743] env[62569]: INFO nova.scheduler.client.report [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Deleted allocations for instance 94392d48-223d-4205-9627-25488a468769 [ 828.265908] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Task: {'id': task-1250148, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.546858} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.265908] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 6e83fd32-215a-42e4-99e7-72bdce25555e/6e83fd32-215a-42e4-99e7-72bdce25555e.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 828.265908] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.265908] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6bc0fc9e-9c11-4069-a74b-69072364e33e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.273175] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Waiting for the task: (returnval){ [ 828.273175] env[62569]: value = "task-1250150" [ 828.273175] env[62569]: _type = "Task" [ 828.273175] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.281606] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Task: {'id': task-1250150, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.422842] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f9840f-e26f-035d-bbb7-09742e45ceda, 'name': SearchDatastore_Task, 'duration_secs': 0.015941} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.423673] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abbaf1a4-a75c-4340-883b-29288e0313fe {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.429357] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 828.429357] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5212551e-f683-e425-8cd2-02e5c5003c36" [ 828.429357] env[62569]: _type = "Task" [ 828.429357] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.438472] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5212551e-f683-e425-8cd2-02e5c5003c36, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.494621] env[62569]: DEBUG oslo_concurrency.lockutils [req-17bccfbd-0bfb-40d5-bda8-89177b8f36c4 req-2f39d5a0-91e3-45ad-a86d-e3c498db62a7 service nova] Releasing lock "refresh_cache-c9568fcb-745f-4c04-8487-a38a617aab7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.495011] env[62569]: DEBUG nova.compute.manager [req-17bccfbd-0bfb-40d5-bda8-89177b8f36c4 req-2f39d5a0-91e3-45ad-a86d-e3c498db62a7 service nova] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Received event network-changed-cbc4229a-f42d-463c-9a46-6a4d6b3581b0 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 828.495218] env[62569]: DEBUG nova.compute.manager [req-17bccfbd-0bfb-40d5-bda8-89177b8f36c4 req-2f39d5a0-91e3-45ad-a86d-e3c498db62a7 service nova] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Refreshing instance network info cache due to event network-changed-cbc4229a-f42d-463c-9a46-6a4d6b3581b0. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 828.495514] env[62569]: DEBUG oslo_concurrency.lockutils [req-17bccfbd-0bfb-40d5-bda8-89177b8f36c4 req-2f39d5a0-91e3-45ad-a86d-e3c498db62a7 service nova] Acquiring lock "refresh_cache-cb9d8991-b7ba-4bf7-9d22-1391f4cea768" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.495616] env[62569]: DEBUG oslo_concurrency.lockutils [req-17bccfbd-0bfb-40d5-bda8-89177b8f36c4 req-2f39d5a0-91e3-45ad-a86d-e3c498db62a7 service nova] Acquired lock "refresh_cache-cb9d8991-b7ba-4bf7-9d22-1391f4cea768" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.495807] env[62569]: DEBUG nova.network.neutron [req-17bccfbd-0bfb-40d5-bda8-89177b8f36c4 req-2f39d5a0-91e3-45ad-a86d-e3c498db62a7 service nova] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Refreshing network info cache for port cbc4229a-f42d-463c-9a46-6a4d6b3581b0 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 828.558077] env[62569]: DEBUG nova.network.neutron [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Updating instance_info_cache with network_info: [{"id": "3da67b46-e89d-4a96-8dd9-752956b9b270", "address": "fa:16:3e:f6:2c:c2", "network": {"id": "ba6b71b1-bd7b-493e-8682-68d842fa935f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1892479706-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e9ea55b00b574a7db0530efcb498ca2b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3da67b46-e8", "ovs_interfaceid": "3da67b46-e89d-4a96-8dd9-752956b9b270", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.569595] env[62569]: DEBUG oslo_vmware.api [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250149, 'name': PowerOnVM_Task, 'duration_secs': 0.972835} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.570457] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 828.570706] env[62569]: INFO nova.compute.manager [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Took 9.22 seconds to spawn the instance on the hypervisor. [ 828.570909] env[62569]: DEBUG nova.compute.manager [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 828.571800] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b48a593-d8a6-4395-a9c4-a9ddc70e1792 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.718284] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2d7fe4cb-1429-4cb1-be70-8cdcf235ab01 tempest-ServerShowV247Test-1047940235 tempest-ServerShowV247Test-1047940235-project-member] Lock "94392d48-223d-4205-9627-25488a468769" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.543s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.783570] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Task: {'id': task-1250150, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094413} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.783570] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 828.784341] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-818d191e-fce6-4d84-8635-be8f4b19a6d9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.811394] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 6e83fd32-215a-42e4-99e7-72bdce25555e/6e83fd32-215a-42e4-99e7-72bdce25555e.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 828.812105] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9cd2659-1ab3-4ec5-bdaa-e65604da3dcc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.831871] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Waiting for the task: (returnval){ [ 828.831871] env[62569]: value = "task-1250151" [ 828.831871] env[62569]: _type = "Task" [ 828.831871] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.839809] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Task: {'id': task-1250151, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.940665] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5212551e-f683-e425-8cd2-02e5c5003c36, 'name': SearchDatastore_Task, 'duration_secs': 0.009653} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.940974] env[62569]: DEBUG oslo_concurrency.lockutils [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.941251] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] c9568fcb-745f-4c04-8487-a38a617aab7c/c9568fcb-745f-4c04-8487-a38a617aab7c.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 828.941514] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b9f6bfc1-84e0-4aff-a3b0-ea8fb2078f38 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.948123] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 828.948123] env[62569]: value = "task-1250152" [ 828.948123] env[62569]: _type = "Task" [ 828.948123] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.955828] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250152, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.061103] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Releasing lock "refresh_cache-a0b406c3-9466-41bd-9de1-e675cab2ceef" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.061449] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Updated the network info_cache for instance {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10379}} [ 829.061752] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 829.062032] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 829.062322] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 829.062573] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 829.062836] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 829.063121] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 829.063339] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62569) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 829.063835] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 829.091284] env[62569]: INFO nova.compute.manager [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Took 35.69 seconds to build instance. [ 829.293812] env[62569]: DEBUG nova.network.neutron [req-17bccfbd-0bfb-40d5-bda8-89177b8f36c4 req-2f39d5a0-91e3-45ad-a86d-e3c498db62a7 service nova] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Updated VIF entry in instance network info cache for port cbc4229a-f42d-463c-9a46-6a4d6b3581b0. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 829.294254] env[62569]: DEBUG nova.network.neutron [req-17bccfbd-0bfb-40d5-bda8-89177b8f36c4 req-2f39d5a0-91e3-45ad-a86d-e3c498db62a7 service nova] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Updating instance_info_cache with network_info: [{"id": "cbc4229a-f42d-463c-9a46-6a4d6b3581b0", "address": "fa:16:3e:14:62:67", "network": {"id": "334a2e22-0249-48e3-a2f5-ac37df9771d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1048653043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f7fb3e698324f8e851aefbee74ac669", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f880ac2e-d532-4f54-87bb-998a8d1bca78", "external-id": "nsx-vlan-transportzone-491", "segmentation_id": 491, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcbc4229a-f4", "ovs_interfaceid": "cbc4229a-f42d-463c-9a46-6a4d6b3581b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.345906] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Task: {'id': task-1250151, 'name': ReconfigVM_Task, 'duration_secs': 0.349287} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.346402] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 6e83fd32-215a-42e4-99e7-72bdce25555e/6e83fd32-215a-42e4-99e7-72bdce25555e.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 829.347374] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f77ea35c-bbfd-4571-828f-8c7d51d45938 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.354743] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Waiting for the task: (returnval){ [ 829.354743] env[62569]: value = "task-1250153" [ 829.354743] env[62569]: _type = "Task" [ 829.354743] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.368125] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Task: {'id': task-1250153, 'name': Rename_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.458714] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250152, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.47115} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.462051] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] c9568fcb-745f-4c04-8487-a38a617aab7c/c9568fcb-745f-4c04-8487-a38a617aab7c.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 829.462051] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 829.462229] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-68974fe8-a7cb-47e1-8579-0e3a253679d7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.468408] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 829.468408] env[62569]: value = "task-1250154" [ 829.468408] env[62569]: _type = "Task" [ 829.468408] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.479290] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250154, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.509416] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1c7923e-4295-4418-82f4-c3b4ef6e285e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.517565] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-954213b4-5bf0-420b-8692-043e81fc2e4e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.550684] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6b67ab4-7979-4b7e-a568-299f478d496a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.558080] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d35f9dd0-3acd-4db4-ba77-146a39c07a63 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.577496] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.578101] env[62569]: DEBUG nova.compute.provider_tree [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.593961] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2decafb8-e312-4f09-ad30-f6a7879fc444 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.861s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.796807] env[62569]: DEBUG oslo_concurrency.lockutils [req-17bccfbd-0bfb-40d5-bda8-89177b8f36c4 req-2f39d5a0-91e3-45ad-a86d-e3c498db62a7 service nova] Releasing lock "refresh_cache-cb9d8991-b7ba-4bf7-9d22-1391f4cea768" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.869941] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Task: {'id': task-1250153, 'name': Rename_Task, 'duration_secs': 0.150293} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.870276] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 829.871329] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2da7e9a6-01d9-4f6a-8b4b-e41d4cf004d6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.879471] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Waiting for the task: (returnval){ [ 829.879471] env[62569]: value = "task-1250155" [ 829.879471] env[62569]: _type = "Task" [ 829.879471] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.894370] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Task: {'id': task-1250155, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.978689] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250154, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063952} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.979219] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 829.980110] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-821022fa-1a5c-4c06-86cb-6a0aa66219d7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.004360] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] c9568fcb-745f-4c04-8487-a38a617aab7c/c9568fcb-745f-4c04-8487-a38a617aab7c.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 830.005567] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-02a39714-0a93-4765-9bff-621bd4674289 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.025707] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 830.025707] env[62569]: value = "task-1250156" [ 830.025707] env[62569]: _type = "Task" [ 830.025707] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.034485] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250156, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.083804] env[62569]: DEBUG nova.scheduler.client.report [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 830.252600] env[62569]: DEBUG nova.compute.manager [req-67b572d9-3f0b-42d1-a3d5-bbaac5f489e3 req-a9b96047-88ae-4638-aa9a-781b1de700f8 service nova] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Received event network-changed-39e71a4f-3b81-4b34-8cc2-2a5352eca4ce {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 830.252880] env[62569]: DEBUG nova.compute.manager [req-67b572d9-3f0b-42d1-a3d5-bbaac5f489e3 req-a9b96047-88ae-4638-aa9a-781b1de700f8 service nova] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Refreshing instance network info cache due to event network-changed-39e71a4f-3b81-4b34-8cc2-2a5352eca4ce. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 830.253166] env[62569]: DEBUG oslo_concurrency.lockutils [req-67b572d9-3f0b-42d1-a3d5-bbaac5f489e3 req-a9b96047-88ae-4638-aa9a-781b1de700f8 service nova] Acquiring lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.253361] env[62569]: DEBUG oslo_concurrency.lockutils [req-67b572d9-3f0b-42d1-a3d5-bbaac5f489e3 req-a9b96047-88ae-4638-aa9a-781b1de700f8 service nova] Acquired lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.253570] env[62569]: DEBUG nova.network.neutron [req-67b572d9-3f0b-42d1-a3d5-bbaac5f489e3 req-a9b96047-88ae-4638-aa9a-781b1de700f8 service nova] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Refreshing network info cache for port 39e71a4f-3b81-4b34-8cc2-2a5352eca4ce {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 830.396025] env[62569]: DEBUG oslo_vmware.api [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Task: {'id': task-1250155, 'name': PowerOnVM_Task, 'duration_secs': 0.459859} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.399514] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 830.399514] env[62569]: INFO nova.compute.manager [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Took 7.55 seconds to spawn the instance on the hypervisor. [ 830.399514] env[62569]: DEBUG nova.compute.manager [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 830.399514] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5729322a-4ece-4049-b1b7-09f5df23757d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.535750] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250156, 'name': ReconfigVM_Task, 'duration_secs': 0.402333} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.536206] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Reconfigured VM instance instance-0000004b to attach disk [datastore1] c9568fcb-745f-4c04-8487-a38a617aab7c/c9568fcb-745f-4c04-8487-a38a617aab7c.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 830.536865] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c47bfe21-8d66-40a7-a1da-26e7bd034505 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.542947] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 830.542947] env[62569]: value = "task-1250157" [ 830.542947] env[62569]: _type = "Task" [ 830.542947] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.550336] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250157, 'name': Rename_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.592206] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.413s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.592206] env[62569]: DEBUG nova.compute.manager [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 830.593288] env[62569]: DEBUG oslo_concurrency.lockutils [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.961s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.593512] env[62569]: DEBUG nova.objects.instance [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Lazy-loading 'resources' on Instance uuid b3ccc28b-68d0-461d-a67d-b5ad179a80f9 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 830.916869] env[62569]: INFO nova.compute.manager [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Took 36.63 seconds to build instance. [ 831.054144] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250157, 'name': Rename_Task, 'duration_secs': 0.453446} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.054503] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 831.054813] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e3e051e7-0c57-492f-8dc2-aed8d7559cb8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.061846] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 831.061846] env[62569]: value = "task-1250158" [ 831.061846] env[62569]: _type = "Task" [ 831.061846] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.068909] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250158, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.089171] env[62569]: DEBUG nova.network.neutron [req-67b572d9-3f0b-42d1-a3d5-bbaac5f489e3 req-a9b96047-88ae-4638-aa9a-781b1de700f8 service nova] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Updated VIF entry in instance network info cache for port 39e71a4f-3b81-4b34-8cc2-2a5352eca4ce. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 831.089311] env[62569]: DEBUG nova.network.neutron [req-67b572d9-3f0b-42d1-a3d5-bbaac5f489e3 req-a9b96047-88ae-4638-aa9a-781b1de700f8 service nova] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Updating instance_info_cache with network_info: [{"id": "39e71a4f-3b81-4b34-8cc2-2a5352eca4ce", "address": "fa:16:3e:4b:e2:a8", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap39e71a4f-3b", "ovs_interfaceid": "39e71a4f-3b81-4b34-8cc2-2a5352eca4ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.102686] env[62569]: DEBUG nova.compute.utils [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 831.107112] env[62569]: DEBUG nova.compute.manager [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 831.107112] env[62569]: DEBUG nova.network.neutron [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 831.194229] env[62569]: DEBUG nova.policy [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '828b6693465e470cbd20c08400767af0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f881ca8a45234848aba4416ff1d582f4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 831.419338] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a4b701e9-9275-4d30-ad8c-a1c2fb05f7c8 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Lock "6e83fd32-215a-42e4-99e7-72bdce25555e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.755s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.447632] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83c7cbdc-16f9-4857-8817-0dcee41b3930 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.457068] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-916218a1-4b81-4cba-a081-b426cb4ab6e0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.492318] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51dcf4d6-9d5c-4d2e-990d-ba9469413983 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.501303] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6729b3d3-1a14-4eb6-b704-2c9a83b1255a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.518310] env[62569]: DEBUG nova.compute.provider_tree [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 831.573172] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250158, 'name': PowerOnVM_Task} progress is 98%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.593159] env[62569]: DEBUG oslo_concurrency.lockutils [req-67b572d9-3f0b-42d1-a3d5-bbaac5f489e3 req-a9b96047-88ae-4638-aa9a-781b1de700f8 service nova] Releasing lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.613619] env[62569]: DEBUG nova.compute.manager [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 831.661117] env[62569]: DEBUG nova.network.neutron [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Successfully created port: 96148207-1e82-4a26-baa5-65f88f889f97 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 831.871497] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Acquiring lock "6e83fd32-215a-42e4-99e7-72bdce25555e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.871791] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Lock "6e83fd32-215a-42e4-99e7-72bdce25555e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.872018] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Acquiring lock "6e83fd32-215a-42e4-99e7-72bdce25555e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.872478] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Lock "6e83fd32-215a-42e4-99e7-72bdce25555e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.873584] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Lock "6e83fd32-215a-42e4-99e7-72bdce25555e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.874918] env[62569]: INFO nova.compute.manager [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Terminating instance [ 832.025041] env[62569]: DEBUG nova.scheduler.client.report [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 832.072161] env[62569]: DEBUG oslo_vmware.api [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250158, 'name': PowerOnVM_Task, 'duration_secs': 0.559363} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.072522] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 832.072800] env[62569]: INFO nova.compute.manager [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Took 7.72 seconds to spawn the instance on the hypervisor. [ 832.073362] env[62569]: DEBUG nova.compute.manager [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 832.074318] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a97902-3365-41ea-b74a-a08a57bb9a5d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.379534] env[62569]: DEBUG nova.compute.manager [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 832.380360] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 832.380844] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1321a0b-75c3-43c9-bfa0-b067a47e302d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.389546] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 832.389806] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-625c4e3b-ec92-409e-bcac-70f7b4927b4f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.395809] env[62569]: DEBUG oslo_vmware.api [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Waiting for the task: (returnval){ [ 832.395809] env[62569]: value = "task-1250159" [ 832.395809] env[62569]: _type = "Task" [ 832.395809] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.404665] env[62569]: DEBUG oslo_vmware.api [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Task: {'id': task-1250159, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.531227] env[62569]: DEBUG oslo_concurrency.lockutils [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.938s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.533567] env[62569]: DEBUG oslo_concurrency.lockutils [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.840s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.535588] env[62569]: DEBUG nova.objects.instance [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lazy-loading 'resources' on Instance uuid 22094c32-5f50-4f86-a77b-cd4adcf8998a {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 832.561590] env[62569]: INFO nova.scheduler.client.report [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Deleted allocations for instance b3ccc28b-68d0-461d-a67d-b5ad179a80f9 [ 832.591408] env[62569]: INFO nova.compute.manager [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Took 35.28 seconds to build instance. [ 832.625844] env[62569]: DEBUG nova.compute.manager [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 832.656159] env[62569]: DEBUG nova.virt.hardware [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 832.656566] env[62569]: DEBUG nova.virt.hardware [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 832.656832] env[62569]: DEBUG nova.virt.hardware [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 832.657653] env[62569]: DEBUG nova.virt.hardware [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 832.657937] env[62569]: DEBUG nova.virt.hardware [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 832.658176] env[62569]: DEBUG nova.virt.hardware [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 832.658475] env[62569]: DEBUG nova.virt.hardware [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 832.658778] env[62569]: DEBUG nova.virt.hardware [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 832.659044] env[62569]: DEBUG nova.virt.hardware [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 832.659273] env[62569]: DEBUG nova.virt.hardware [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 832.659497] env[62569]: DEBUG nova.virt.hardware [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 832.660444] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96d7fa6b-14f9-44a2-9334-b84f0ffa6d6f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.670844] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b45cbe8-4887-482d-bccd-64d03d70f374 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.911405] env[62569]: DEBUG oslo_vmware.api [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Task: {'id': task-1250159, 'name': PowerOffVM_Task, 'duration_secs': 0.352615} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.911405] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 832.911405] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 832.911674] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9ec5d65e-fd0e-4cc3-91ce-7f408fdd5c4d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.977259] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 832.977259] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 832.977381] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Deleting the datastore file [datastore1] 6e83fd32-215a-42e4-99e7-72bdce25555e {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 832.977658] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-62327f76-73fc-4326-9f45-52cb2ffec89e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.984628] env[62569]: DEBUG oslo_vmware.api [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Waiting for the task: (returnval){ [ 832.984628] env[62569]: value = "task-1250161" [ 832.984628] env[62569]: _type = "Task" [ 832.984628] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.996696] env[62569]: DEBUG oslo_vmware.api [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Task: {'id': task-1250161, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.076465] env[62569]: DEBUG oslo_concurrency.lockutils [None req-630e4c87-92da-436e-a7c3-c8a1e6ffd9c8 tempest-FloatingIPsAssociationNegativeTestJSON-833233566 tempest-FloatingIPsAssociationNegativeTestJSON-833233566-project-member] Lock "b3ccc28b-68d0-461d-a67d-b5ad179a80f9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.476s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.093698] env[62569]: DEBUG oslo_concurrency.lockutils [None req-176069a3-7b45-4416-a4dd-f04c066425c0 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "c9568fcb-745f-4c04-8487-a38a617aab7c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.456s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.324129] env[62569]: DEBUG nova.compute.manager [req-61f75dae-d7cb-4f0b-b5bc-50a791052092 req-ef0d281b-bc3c-4371-878f-589139f886ff service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Received event network-vif-plugged-96148207-1e82-4a26-baa5-65f88f889f97 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 833.324797] env[62569]: DEBUG oslo_concurrency.lockutils [req-61f75dae-d7cb-4f0b-b5bc-50a791052092 req-ef0d281b-bc3c-4371-878f-589139f886ff service nova] Acquiring lock "a5b3c345-149b-4dc3-84ff-fe72183efadb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.324797] env[62569]: DEBUG oslo_concurrency.lockutils [req-61f75dae-d7cb-4f0b-b5bc-50a791052092 req-ef0d281b-bc3c-4371-878f-589139f886ff service nova] Lock "a5b3c345-149b-4dc3-84ff-fe72183efadb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.324797] env[62569]: DEBUG oslo_concurrency.lockutils [req-61f75dae-d7cb-4f0b-b5bc-50a791052092 req-ef0d281b-bc3c-4371-878f-589139f886ff service nova] Lock "a5b3c345-149b-4dc3-84ff-fe72183efadb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.325212] env[62569]: DEBUG nova.compute.manager [req-61f75dae-d7cb-4f0b-b5bc-50a791052092 req-ef0d281b-bc3c-4371-878f-589139f886ff service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] No waiting events found dispatching network-vif-plugged-96148207-1e82-4a26-baa5-65f88f889f97 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 833.325427] env[62569]: WARNING nova.compute.manager [req-61f75dae-d7cb-4f0b-b5bc-50a791052092 req-ef0d281b-bc3c-4371-878f-589139f886ff service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Received unexpected event network-vif-plugged-96148207-1e82-4a26-baa5-65f88f889f97 for instance with vm_state building and task_state spawning. [ 833.412646] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44e53822-9dbc-46f7-be48-878abdb118ed {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.421746] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b31354aa-2987-4737-9e91-5cb317999b9b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.456640] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1488cc0c-e081-43fc-9590-83cabbb61257 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.465305] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-409e3f22-6e94-4cfb-9fce-4e6ab54082fc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.480615] env[62569]: DEBUG nova.compute.provider_tree [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 833.496968] env[62569]: DEBUG oslo_concurrency.lockutils [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Acquiring lock "691d8fbe-b9ee-454f-bd7b-14520e53ed26" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.497723] env[62569]: DEBUG oslo_concurrency.lockutils [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Lock "691d8fbe-b9ee-454f-bd7b-14520e53ed26" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.501330] env[62569]: DEBUG nova.network.neutron [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Successfully updated port: 96148207-1e82-4a26-baa5-65f88f889f97 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 833.512408] env[62569]: DEBUG oslo_vmware.api [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Task: {'id': task-1250161, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.426954} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.512757] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 833.512984] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 833.513205] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 833.513382] env[62569]: INFO nova.compute.manager [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Took 1.13 seconds to destroy the instance on the hypervisor. [ 833.513628] env[62569]: DEBUG oslo.service.loopingcall [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.515526] env[62569]: DEBUG nova.compute.manager [-] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 833.515526] env[62569]: DEBUG nova.network.neutron [-] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 833.990024] env[62569]: DEBUG nova.scheduler.client.report [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 834.000779] env[62569]: DEBUG nova.compute.manager [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 834.006919] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "c9568fcb-745f-4c04-8487-a38a617aab7c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.007358] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "c9568fcb-745f-4c04-8487-a38a617aab7c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.007535] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "c9568fcb-745f-4c04-8487-a38a617aab7c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.007745] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "c9568fcb-745f-4c04-8487-a38a617aab7c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.007936] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "c9568fcb-745f-4c04-8487-a38a617aab7c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.010169] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Acquiring lock "refresh_cache-a5b3c345-149b-4dc3-84ff-fe72183efadb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.010306] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Acquired lock "refresh_cache-a5b3c345-149b-4dc3-84ff-fe72183efadb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.010452] env[62569]: DEBUG nova.network.neutron [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 834.012805] env[62569]: INFO nova.compute.manager [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Terminating instance [ 834.346662] env[62569]: DEBUG nova.network.neutron [-] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.495275] env[62569]: DEBUG oslo_concurrency.lockutils [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.962s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.498693] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.195s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.502806] env[62569]: INFO nova.compute.claims [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 834.520577] env[62569]: DEBUG nova.compute.manager [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 834.522617] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 834.522617] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8deeedc9-164a-47fb-9756-81daeda852ef {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.533279] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 834.536290] env[62569]: INFO nova.scheduler.client.report [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Deleted allocations for instance 22094c32-5f50-4f86-a77b-cd4adcf8998a [ 834.540021] env[62569]: DEBUG oslo_concurrency.lockutils [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.540021] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-426e3b1f-5c3a-4660-afc4-5ee9212a627a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.551465] env[62569]: DEBUG oslo_vmware.api [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 834.551465] env[62569]: value = "task-1250162" [ 834.551465] env[62569]: _type = "Task" [ 834.551465] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.572753] env[62569]: DEBUG oslo_vmware.api [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250162, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.611523] env[62569]: DEBUG nova.network.neutron [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 834.852166] env[62569]: INFO nova.compute.manager [-] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Took 1.34 seconds to deallocate network for instance. [ 834.940694] env[62569]: DEBUG nova.network.neutron [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Updating instance_info_cache with network_info: [{"id": "96148207-1e82-4a26-baa5-65f88f889f97", "address": "fa:16:3e:03:33:1e", "network": {"id": "68456414-139e-4c9b-928c-1eba44e74d7f", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1573053904-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f881ca8a45234848aba4416ff1d582f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96148207-1e", "ovs_interfaceid": "96148207-1e82-4a26-baa5-65f88f889f97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.050588] env[62569]: DEBUG oslo_concurrency.lockutils [None req-506822e2-4868-4e35-ba64-7cbbf2d578fb tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "22094c32-5f50-4f86-a77b-cd4adcf8998a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.855s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.069916] env[62569]: DEBUG oslo_vmware.api [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250162, 'name': PowerOffVM_Task, 'duration_secs': 0.429841} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.070362] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 835.070667] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 835.071046] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-564382fd-a689-4cde-a04d-6992621199b8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.363493] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.367810] env[62569]: DEBUG oslo_concurrency.lockutils [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquiring lock "337d88a2-30b9-4846-929e-042bd7a64a65" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.367810] env[62569]: DEBUG oslo_concurrency.lockutils [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "337d88a2-30b9-4846-929e-042bd7a64a65" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.367810] env[62569]: INFO nova.compute.manager [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Shelving [ 835.443999] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Releasing lock "refresh_cache-a5b3c345-149b-4dc3-84ff-fe72183efadb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.444367] env[62569]: DEBUG nova.compute.manager [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Instance network_info: |[{"id": "96148207-1e82-4a26-baa5-65f88f889f97", "address": "fa:16:3e:03:33:1e", "network": {"id": "68456414-139e-4c9b-928c-1eba44e74d7f", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1573053904-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f881ca8a45234848aba4416ff1d582f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96148207-1e", "ovs_interfaceid": "96148207-1e82-4a26-baa5-65f88f889f97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 835.445306] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:33:1e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '68add7d6-c025-46fa-84d3-9c589adb63e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '96148207-1e82-4a26-baa5-65f88f889f97', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 835.454514] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Creating folder: Project (f881ca8a45234848aba4416ff1d582f4). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 835.459134] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-58577af4-8423-4db4-9e97-f666102c310c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.461698] env[62569]: DEBUG nova.compute.manager [req-4f7efbec-305d-4c86-997a-2a9b2fb90e32 req-2bff1307-dc8e-4ee8-addd-5b09df4be3f7 service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Received event network-changed-96148207-1e82-4a26-baa5-65f88f889f97 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 835.461953] env[62569]: DEBUG nova.compute.manager [req-4f7efbec-305d-4c86-997a-2a9b2fb90e32 req-2bff1307-dc8e-4ee8-addd-5b09df4be3f7 service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Refreshing instance network info cache due to event network-changed-96148207-1e82-4a26-baa5-65f88f889f97. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 835.462153] env[62569]: DEBUG oslo_concurrency.lockutils [req-4f7efbec-305d-4c86-997a-2a9b2fb90e32 req-2bff1307-dc8e-4ee8-addd-5b09df4be3f7 service nova] Acquiring lock "refresh_cache-a5b3c345-149b-4dc3-84ff-fe72183efadb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.462353] env[62569]: DEBUG oslo_concurrency.lockutils [req-4f7efbec-305d-4c86-997a-2a9b2fb90e32 req-2bff1307-dc8e-4ee8-addd-5b09df4be3f7 service nova] Acquired lock "refresh_cache-a5b3c345-149b-4dc3-84ff-fe72183efadb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.462599] env[62569]: DEBUG nova.network.neutron [req-4f7efbec-305d-4c86-997a-2a9b2fb90e32 req-2bff1307-dc8e-4ee8-addd-5b09df4be3f7 service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Refreshing network info cache for port 96148207-1e82-4a26-baa5-65f88f889f97 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 835.477380] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Created folder: Project (f881ca8a45234848aba4416ff1d582f4) in parent group-v269330. [ 835.477553] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Creating folder: Instances. Parent ref: group-v269428. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 835.477842] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-61a4c846-2c65-41eb-aa2e-db543e306f24 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.488038] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Created folder: Instances in parent group-v269428. [ 835.489519] env[62569]: DEBUG oslo.service.loopingcall [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 835.490600] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 835.491216] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-08d01246-834c-443b-ac2e-047e068e2f55 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.513345] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 835.513345] env[62569]: value = "task-1250166" [ 835.513345] env[62569]: _type = "Task" [ 835.513345] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.525144] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250166, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.767814] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f60e813f-3eb8-4e22-ac68-e347a0691c11 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.775334] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b5c18e5-7b9b-4808-b69a-fa5ad757b9f4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.807107] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b99128e9-ed12-41d2-90b8-01593aae46e6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.814383] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262b9c1d-d8ed-4f90-a1f6-9231a39d2325 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.827831] env[62569]: DEBUG nova.compute.provider_tree [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 836.024472] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250166, 'name': CreateVM_Task, 'duration_secs': 0.313328} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.024703] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 836.026906] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.028082] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.028263] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 836.029064] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41744580-3746-4039-b12c-ab6ff063e4f8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.036809] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Waiting for the task: (returnval){ [ 836.036809] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52acfeac-586e-03c0-b7f4-fcf02970d4db" [ 836.036809] env[62569]: _type = "Task" [ 836.036809] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.047432] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52acfeac-586e-03c0-b7f4-fcf02970d4db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.185455] env[62569]: DEBUG nova.network.neutron [req-4f7efbec-305d-4c86-997a-2a9b2fb90e32 req-2bff1307-dc8e-4ee8-addd-5b09df4be3f7 service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Updated VIF entry in instance network info cache for port 96148207-1e82-4a26-baa5-65f88f889f97. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 836.185842] env[62569]: DEBUG nova.network.neutron [req-4f7efbec-305d-4c86-997a-2a9b2fb90e32 req-2bff1307-dc8e-4ee8-addd-5b09df4be3f7 service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Updating instance_info_cache with network_info: [{"id": "96148207-1e82-4a26-baa5-65f88f889f97", "address": "fa:16:3e:03:33:1e", "network": {"id": "68456414-139e-4c9b-928c-1eba44e74d7f", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1573053904-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f881ca8a45234848aba4416ff1d582f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96148207-1e", "ovs_interfaceid": "96148207-1e82-4a26-baa5-65f88f889f97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.331346] env[62569]: DEBUG nova.scheduler.client.report [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 836.376023] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 836.376371] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8b259a36-d03f-46a9-a55f-6522fecfd90f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.384529] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 836.384529] env[62569]: value = "task-1250167" [ 836.384529] env[62569]: _type = "Task" [ 836.384529] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.393087] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250167, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.548483] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52acfeac-586e-03c0-b7f4-fcf02970d4db, 'name': SearchDatastore_Task, 'duration_secs': 0.009973} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.548855] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.549116] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 836.549434] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.549604] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.549808] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 836.550119] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3040cbc9-b8ec-4626-b97f-a70e32f4f8ee {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.558776] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 836.558966] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 836.559731] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-adcd01f3-ebff-41af-b5ab-71b13b10b858 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.564982] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Waiting for the task: (returnval){ [ 836.564982] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ff93f4-738e-f2a5-5284-913694adf1ca" [ 836.564982] env[62569]: _type = "Task" [ 836.564982] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.572662] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ff93f4-738e-f2a5-5284-913694adf1ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.688866] env[62569]: DEBUG oslo_concurrency.lockutils [req-4f7efbec-305d-4c86-997a-2a9b2fb90e32 req-2bff1307-dc8e-4ee8-addd-5b09df4be3f7 service nova] Releasing lock "refresh_cache-a5b3c345-149b-4dc3-84ff-fe72183efadb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.689638] env[62569]: DEBUG nova.compute.manager [req-4f7efbec-305d-4c86-997a-2a9b2fb90e32 req-2bff1307-dc8e-4ee8-addd-5b09df4be3f7 service nova] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Received event network-vif-deleted-8e0244f3-9b3e-4fe8-8cfd-666a36f1ef58 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 836.840018] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.339s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.840018] env[62569]: DEBUG nova.compute.manager [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 836.840346] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 23.992s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.840686] env[62569]: DEBUG nova.objects.instance [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62569) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 836.896136] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250167, 'name': PowerOffVM_Task, 'duration_secs': 0.219187} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.896136] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 836.896545] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89c6291-368f-44bb-9e90-3881e43a4953 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.926128] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99976e3e-68c2-405a-9250-dc30075059c5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.076204] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ff93f4-738e-f2a5-5284-913694adf1ca, 'name': SearchDatastore_Task, 'duration_secs': 0.008501} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.077030] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7acec72-5eb6-4f91-abde-2bb4370d975c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.082428] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Waiting for the task: (returnval){ [ 837.082428] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52bfceba-1514-284d-9a8b-8ffdee0f9a79" [ 837.082428] env[62569]: _type = "Task" [ 837.082428] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.093036] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52bfceba-1514-284d-9a8b-8ffdee0f9a79, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.344942] env[62569]: DEBUG nova.compute.utils [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 837.349505] env[62569]: DEBUG nova.compute.manager [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 837.350487] env[62569]: DEBUG nova.network.neutron [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 837.409067] env[62569]: DEBUG nova.policy [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a272992a4894805bfb958680a37da8e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b20340a1ce0447cae1bfd8b7c28928f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 837.438331] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Creating Snapshot of the VM instance {{(pid=62569) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 837.438849] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-23817aed-c3dd-4651-97f8-3d3ecac0654c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.445947] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 837.446203] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 837.446391] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Deleting the datastore file [datastore1] c9568fcb-745f-4c04-8487-a38a617aab7c {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 837.446665] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cc685083-1c7a-4da0-8fb9-d8ac6aa3c071 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.449834] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 837.449834] env[62569]: value = "task-1250168" [ 837.449834] env[62569]: _type = "Task" [ 837.449834] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.455031] env[62569]: DEBUG oslo_vmware.api [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 837.455031] env[62569]: value = "task-1250169" [ 837.455031] env[62569]: _type = "Task" [ 837.455031] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.460578] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250168, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.466507] env[62569]: DEBUG oslo_vmware.api [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250169, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.597107] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52bfceba-1514-284d-9a8b-8ffdee0f9a79, 'name': SearchDatastore_Task, 'duration_secs': 0.009655} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.597439] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.597716] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] a5b3c345-149b-4dc3-84ff-fe72183efadb/a5b3c345-149b-4dc3-84ff-fe72183efadb.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 837.598011] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f35f19fd-7645-4d0b-8938-e8dabe683b58 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.604442] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Waiting for the task: (returnval){ [ 837.604442] env[62569]: value = "task-1250170" [ 837.604442] env[62569]: _type = "Task" [ 837.604442] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.613342] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250170, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.853116] env[62569]: DEBUG nova.compute.manager [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 837.855489] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ab854494-3530-4482-ae24-9a3bd39184bd tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.857580] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.322s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.859351] env[62569]: INFO nova.compute.claims [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 837.962974] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250168, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.969884] env[62569]: DEBUG oslo_vmware.api [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250169, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141946} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.970331] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 837.970570] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 837.970937] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 837.971136] env[62569]: INFO nova.compute.manager [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Took 3.45 seconds to destroy the instance on the hypervisor. [ 837.971408] env[62569]: DEBUG oslo.service.loopingcall [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.971603] env[62569]: DEBUG nova.compute.manager [-] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 837.971766] env[62569]: DEBUG nova.network.neutron [-] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 838.117400] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250170, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.460364} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.117670] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] a5b3c345-149b-4dc3-84ff-fe72183efadb/a5b3c345-149b-4dc3-84ff-fe72183efadb.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 838.117883] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 838.120226] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6213320c-41da-48ab-809b-11743da7293e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.132616] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Waiting for the task: (returnval){ [ 838.132616] env[62569]: value = "task-1250171" [ 838.132616] env[62569]: _type = "Task" [ 838.132616] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.143852] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250171, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.157588] env[62569]: DEBUG nova.network.neutron [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Successfully created port: 3d532085-b731-4296-adce-01e80852415b {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 838.466307] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250168, 'name': CreateSnapshot_Task, 'duration_secs': 0.563756} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.467698] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Created Snapshot of the VM instance {{(pid=62569) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 838.469014] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63c4d812-1909-4a2d-a86d-6c7ddf8e32ac {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.644302] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250171, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.098952} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.644657] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 838.645506] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15a18362-0c36-462b-8021-55819f51bc8b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.677886] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] a5b3c345-149b-4dc3-84ff-fe72183efadb/a5b3c345-149b-4dc3-84ff-fe72183efadb.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.679252] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d38e5d4-5294-4e3e-be2c-f7c0137a6432 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.698099] env[62569]: DEBUG nova.compute.manager [req-0cdc20ea-d62a-432a-a4f5-a4bfdef42db5 req-b8ed2337-e0c5-4e0c-8265-e000f884c604 service nova] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Received event network-vif-deleted-174411d8-9538-4849-b0d1-fdb27fe71a7b {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 838.698792] env[62569]: INFO nova.compute.manager [req-0cdc20ea-d62a-432a-a4f5-a4bfdef42db5 req-b8ed2337-e0c5-4e0c-8265-e000f884c604 service nova] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Neutron deleted interface 174411d8-9538-4849-b0d1-fdb27fe71a7b; detaching it from the instance and deleting it from the info cache [ 838.698950] env[62569]: DEBUG nova.network.neutron [req-0cdc20ea-d62a-432a-a4f5-a4bfdef42db5 req-b8ed2337-e0c5-4e0c-8265-e000f884c604 service nova] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.708795] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Waiting for the task: (returnval){ [ 838.708795] env[62569]: value = "task-1250172" [ 838.708795] env[62569]: _type = "Task" [ 838.708795] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.720351] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250172, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.862540] env[62569]: DEBUG nova.compute.manager [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 838.896521] env[62569]: DEBUG nova.virt.hardware [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 838.896790] env[62569]: DEBUG nova.virt.hardware [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 838.897055] env[62569]: DEBUG nova.virt.hardware [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 838.897275] env[62569]: DEBUG nova.virt.hardware [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 838.897349] env[62569]: DEBUG nova.virt.hardware [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 838.897453] env[62569]: DEBUG nova.virt.hardware [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 838.897664] env[62569]: DEBUG nova.virt.hardware [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 838.897831] env[62569]: DEBUG nova.virt.hardware [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 838.898015] env[62569]: DEBUG nova.virt.hardware [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 838.898376] env[62569]: DEBUG nova.virt.hardware [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 838.898609] env[62569]: DEBUG nova.virt.hardware [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 838.899527] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e10810-1165-49bb-8c77-365f9191a137 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.911290] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ced548d-9ef9-48f6-824d-68185702f7fb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.993277] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Creating linked-clone VM from snapshot {{(pid=62569) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 838.996233] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-55279500-c2b8-4e64-afee-aee7fbaa2b6a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.003729] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 839.003729] env[62569]: value = "task-1250173" [ 839.003729] env[62569]: _type = "Task" [ 839.003729] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.016189] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250173, 'name': CloneVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.098488] env[62569]: DEBUG nova.network.neutron [-] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.178407] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01891102-822c-4ba4-b36b-fb8617fc404f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.186577] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cabe78d-309e-45cf-ab75-293dcfe05f2b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.217224] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d995bb67-0579-4dde-8d01-416ceba7506b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.222642] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39ae41dc-c5e5-4972-8e24-3306135e554e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.234739] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2600e12-52a4-4fbc-8f9d-11278189289c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.238932] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250172, 'name': ReconfigVM_Task, 'duration_secs': 0.271993} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.242149] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17151b3a-7cf4-4810-b701-4f3207da0815 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.253646] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Reconfigured VM instance instance-0000004c to attach disk [datastore2] a5b3c345-149b-4dc3-84ff-fe72183efadb/a5b3c345-149b-4dc3-84ff-fe72183efadb.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.255735] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-407b7ddb-ea71-413e-9634-01bfdd1c234b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.268425] env[62569]: DEBUG nova.compute.provider_tree [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.273053] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Waiting for the task: (returnval){ [ 839.273053] env[62569]: value = "task-1250174" [ 839.273053] env[62569]: _type = "Task" [ 839.273053] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.287297] env[62569]: DEBUG nova.compute.manager [req-0cdc20ea-d62a-432a-a4f5-a4bfdef42db5 req-b8ed2337-e0c5-4e0c-8265-e000f884c604 service nova] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Detach interface failed, port_id=174411d8-9538-4849-b0d1-fdb27fe71a7b, reason: Instance c9568fcb-745f-4c04-8487-a38a617aab7c could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 839.293131] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250174, 'name': Rename_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.513635] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250173, 'name': CloneVM_Task} progress is 94%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.601099] env[62569]: INFO nova.compute.manager [-] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Took 1.63 seconds to deallocate network for instance. [ 839.770032] env[62569]: DEBUG nova.scheduler.client.report [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 839.785941] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250174, 'name': Rename_Task, 'duration_secs': 0.147454} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.786388] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 839.786737] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cb9e308d-32da-45b9-b591-c7e41bb88b1a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.796247] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Waiting for the task: (returnval){ [ 839.796247] env[62569]: value = "task-1250175" [ 839.796247] env[62569]: _type = "Task" [ 839.796247] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.803979] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250175, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.016756] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250173, 'name': CloneVM_Task} progress is 94%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.073710] env[62569]: DEBUG nova.network.neutron [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Successfully updated port: 3d532085-b731-4296-adce-01e80852415b {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 840.108443] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.279887] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.422s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.280460] env[62569]: DEBUG nova.compute.manager [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 840.283555] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.307s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.283787] env[62569]: DEBUG nova.objects.instance [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Lazy-loading 'resources' on Instance uuid cefbca2e-609d-4954-bec6-52ffe095446f {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 840.306759] env[62569]: DEBUG oslo_vmware.api [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250175, 'name': PowerOnVM_Task, 'duration_secs': 0.471275} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.307036] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 840.307975] env[62569]: INFO nova.compute.manager [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Took 7.68 seconds to spawn the instance on the hypervisor. [ 840.307975] env[62569]: DEBUG nova.compute.manager [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 840.308237] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-686fd41e-0521-40b4-9955-eda2ad1d10e8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.518834] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250173, 'name': CloneVM_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.577224] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "refresh_cache-05db90a6-3faf-4878-b782-ed17df47ed5f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.577381] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquired lock "refresh_cache-05db90a6-3faf-4878-b782-ed17df47ed5f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.577534] env[62569]: DEBUG nova.network.neutron [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 840.703047] env[62569]: DEBUG nova.compute.manager [req-9a4ed449-afe2-421a-9b74-d5d94d3dfb60 req-cbf68f42-6706-4c87-827b-ae789d541497 service nova] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Received event network-vif-plugged-3d532085-b731-4296-adce-01e80852415b {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 840.703328] env[62569]: DEBUG oslo_concurrency.lockutils [req-9a4ed449-afe2-421a-9b74-d5d94d3dfb60 req-cbf68f42-6706-4c87-827b-ae789d541497 service nova] Acquiring lock "05db90a6-3faf-4878-b782-ed17df47ed5f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.703551] env[62569]: DEBUG oslo_concurrency.lockutils [req-9a4ed449-afe2-421a-9b74-d5d94d3dfb60 req-cbf68f42-6706-4c87-827b-ae789d541497 service nova] Lock "05db90a6-3faf-4878-b782-ed17df47ed5f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.703780] env[62569]: DEBUG oslo_concurrency.lockutils [req-9a4ed449-afe2-421a-9b74-d5d94d3dfb60 req-cbf68f42-6706-4c87-827b-ae789d541497 service nova] Lock "05db90a6-3faf-4878-b782-ed17df47ed5f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.704100] env[62569]: DEBUG nova.compute.manager [req-9a4ed449-afe2-421a-9b74-d5d94d3dfb60 req-cbf68f42-6706-4c87-827b-ae789d541497 service nova] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] No waiting events found dispatching network-vif-plugged-3d532085-b731-4296-adce-01e80852415b {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 840.704293] env[62569]: WARNING nova.compute.manager [req-9a4ed449-afe2-421a-9b74-d5d94d3dfb60 req-cbf68f42-6706-4c87-827b-ae789d541497 service nova] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Received unexpected event network-vif-plugged-3d532085-b731-4296-adce-01e80852415b for instance with vm_state building and task_state spawning. [ 840.704465] env[62569]: DEBUG nova.compute.manager [req-9a4ed449-afe2-421a-9b74-d5d94d3dfb60 req-cbf68f42-6706-4c87-827b-ae789d541497 service nova] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Received event network-changed-3d532085-b731-4296-adce-01e80852415b {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 840.704627] env[62569]: DEBUG nova.compute.manager [req-9a4ed449-afe2-421a-9b74-d5d94d3dfb60 req-cbf68f42-6706-4c87-827b-ae789d541497 service nova] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Refreshing instance network info cache due to event network-changed-3d532085-b731-4296-adce-01e80852415b. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 840.704801] env[62569]: DEBUG oslo_concurrency.lockutils [req-9a4ed449-afe2-421a-9b74-d5d94d3dfb60 req-cbf68f42-6706-4c87-827b-ae789d541497 service nova] Acquiring lock "refresh_cache-05db90a6-3faf-4878-b782-ed17df47ed5f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.791250] env[62569]: DEBUG nova.compute.utils [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 840.795218] env[62569]: DEBUG nova.compute.manager [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 840.798017] env[62569]: DEBUG nova.network.neutron [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 840.836147] env[62569]: INFO nova.compute.manager [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Took 38.79 seconds to build instance. [ 840.852378] env[62569]: DEBUG nova.policy [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5da4f6b7a6784a73bd3fed04b275041b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bdba6022e3d4697a336ca28ca4eccec', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 841.019208] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250173, 'name': CloneVM_Task, 'duration_secs': 1.637868} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.019494] env[62569]: INFO nova.virt.vmwareapi.vmops [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Created linked-clone VM from snapshot [ 841.020286] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c266bab7-682b-4b9f-977c-da28f6ab871e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.031049] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Uploading image 45f3d90c-b69b-4905-993e-439633bfeb29 {{(pid=62569) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 841.058927] env[62569]: DEBUG oslo_vmware.rw_handles [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 841.058927] env[62569]: value = "vm-269432" [ 841.058927] env[62569]: _type = "VirtualMachine" [ 841.058927] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 841.059280] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-2a30f24d-95a9-4f0d-ade5-5452e7899fbe {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.070936] env[62569]: DEBUG oslo_vmware.rw_handles [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lease: (returnval){ [ 841.070936] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f5619b-412d-1a71-467b-e69187b51f60" [ 841.070936] env[62569]: _type = "HttpNfcLease" [ 841.070936] env[62569]: } obtained for exporting VM: (result){ [ 841.070936] env[62569]: value = "vm-269432" [ 841.070936] env[62569]: _type = "VirtualMachine" [ 841.070936] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 841.071214] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the lease: (returnval){ [ 841.071214] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f5619b-412d-1a71-467b-e69187b51f60" [ 841.071214] env[62569]: _type = "HttpNfcLease" [ 841.071214] env[62569]: } to be ready. {{(pid=62569) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 841.087479] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 841.087479] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f5619b-412d-1a71-467b-e69187b51f60" [ 841.087479] env[62569]: _type = "HttpNfcLease" [ 841.087479] env[62569]: } is ready. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 841.087816] env[62569]: DEBUG oslo_vmware.rw_handles [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 841.087816] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f5619b-412d-1a71-467b-e69187b51f60" [ 841.087816] env[62569]: _type = "HttpNfcLease" [ 841.087816] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 841.088620] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e26d8eaf-8555-4248-be4a-a5ddf6ebf901 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.102175] env[62569]: DEBUG oslo_vmware.rw_handles [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bb5346-cfae-5dda-f4b1-7af62ff83a9f/disk-0.vmdk from lease info. {{(pid=62569) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 841.102175] env[62569]: DEBUG oslo_vmware.rw_handles [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bb5346-cfae-5dda-f4b1-7af62ff83a9f/disk-0.vmdk for reading. {{(pid=62569) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 841.166991] env[62569]: DEBUG nova.network.neutron [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 841.170382] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3ccce20-b388-4e90-a20d-47ab46c1fa62 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.180166] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-238c8192-4133-4b42-884c-ba3e2e6c78bd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.184619] env[62569]: DEBUG nova.network.neutron [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Successfully created port: 030d3c5e-456b-4a2d-a229-29d169db3996 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 841.219162] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5932892b-3a22-47f5-8715-b98cfa6cd460 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.222226] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2125a967-8403-4058-acc4-2ff8b17814c8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.231787] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed0237b2-ab3a-4795-bc32-cee5a2e7010a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.248672] env[62569]: DEBUG nova.compute.provider_tree [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 841.298448] env[62569]: DEBUG nova.compute.manager [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 841.340018] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02a7fba2-df5d-4ee2-891b-70f7f1433e1d tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Lock "a5b3c345-149b-4dc3-84ff-fe72183efadb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.073s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.418671] env[62569]: DEBUG nova.network.neutron [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Updating instance_info_cache with network_info: [{"id": "3d532085-b731-4296-adce-01e80852415b", "address": "fa:16:3e:66:17:5b", "network": {"id": "8ea5e138-813b-4c5c-88a1-b1ea79807f10", "bridge": "br-int", "label": "tempest-ImagesTestJSON-524163445-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b20340a1ce0447cae1bfd8b7c28928f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d532085-b7", "ovs_interfaceid": "3d532085-b731-4296-adce-01e80852415b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.754527] env[62569]: DEBUG nova.scheduler.client.report [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 841.860773] env[62569]: INFO nova.compute.manager [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Rescuing [ 841.861174] env[62569]: DEBUG oslo_concurrency.lockutils [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Acquiring lock "refresh_cache-a5b3c345-149b-4dc3-84ff-fe72183efadb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.861381] env[62569]: DEBUG oslo_concurrency.lockutils [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Acquired lock "refresh_cache-a5b3c345-149b-4dc3-84ff-fe72183efadb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.861725] env[62569]: DEBUG nova.network.neutron [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 841.921770] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Releasing lock "refresh_cache-05db90a6-3faf-4878-b782-ed17df47ed5f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.922290] env[62569]: DEBUG nova.compute.manager [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Instance network_info: |[{"id": "3d532085-b731-4296-adce-01e80852415b", "address": "fa:16:3e:66:17:5b", "network": {"id": "8ea5e138-813b-4c5c-88a1-b1ea79807f10", "bridge": "br-int", "label": "tempest-ImagesTestJSON-524163445-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b20340a1ce0447cae1bfd8b7c28928f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d532085-b7", "ovs_interfaceid": "3d532085-b731-4296-adce-01e80852415b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 841.922709] env[62569]: DEBUG oslo_concurrency.lockutils [req-9a4ed449-afe2-421a-9b74-d5d94d3dfb60 req-cbf68f42-6706-4c87-827b-ae789d541497 service nova] Acquired lock "refresh_cache-05db90a6-3faf-4878-b782-ed17df47ed5f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.922930] env[62569]: DEBUG nova.network.neutron [req-9a4ed449-afe2-421a-9b74-d5d94d3dfb60 req-cbf68f42-6706-4c87-827b-ae789d541497 service nova] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Refreshing network info cache for port 3d532085-b731-4296-adce-01e80852415b {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 841.925495] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:17:5b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f65996a3-f865-4492-9377-cd14ec8b3aae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3d532085-b731-4296-adce-01e80852415b', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 841.934482] env[62569]: DEBUG oslo.service.loopingcall [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.935936] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 841.936242] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c62a7cb0-c768-4ef1-8a55-6272d0ab1edc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.959198] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 841.959198] env[62569]: value = "task-1250177" [ 841.959198] env[62569]: _type = "Task" [ 841.959198] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.972916] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250177, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.270711] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.984s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.270711] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.836s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.270973] env[62569]: DEBUG nova.objects.instance [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Lazy-loading 'resources' on Instance uuid 77a1b192-6aff-4fee-93d7-57cebcdce626 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 842.303366] env[62569]: INFO nova.scheduler.client.report [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Deleted allocations for instance cefbca2e-609d-4954-bec6-52ffe095446f [ 842.307188] env[62569]: DEBUG nova.compute.manager [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 842.338886] env[62569]: DEBUG nova.virt.hardware [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 842.340318] env[62569]: DEBUG nova.virt.hardware [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 842.340515] env[62569]: DEBUG nova.virt.hardware [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 842.340780] env[62569]: DEBUG nova.virt.hardware [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 842.340981] env[62569]: DEBUG nova.virt.hardware [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 842.341238] env[62569]: DEBUG nova.virt.hardware [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 842.341549] env[62569]: DEBUG nova.virt.hardware [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 842.341756] env[62569]: DEBUG nova.virt.hardware [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 842.342065] env[62569]: DEBUG nova.virt.hardware [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 842.342351] env[62569]: DEBUG nova.virt.hardware [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 842.342591] env[62569]: DEBUG nova.virt.hardware [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 842.344601] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51e9afd2-ae55-4da4-9847-0f7c69a93191 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.355037] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f3c115a-a69e-443d-a33e-b5b595664397 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.470250] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250177, 'name': CreateVM_Task, 'duration_secs': 0.400072} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.470516] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 842.471321] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.471593] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.472188] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 842.474740] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a55958a1-f63e-4500-93aa-af8ea8fabd9c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.481062] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 842.481062] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52844c82-d367-1d31-8e22-61083e68d1e7" [ 842.481062] env[62569]: _type = "Task" [ 842.481062] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.490528] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52844c82-d367-1d31-8e22-61083e68d1e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.664994] env[62569]: DEBUG nova.network.neutron [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Updating instance_info_cache with network_info: [{"id": "96148207-1e82-4a26-baa5-65f88f889f97", "address": "fa:16:3e:03:33:1e", "network": {"id": "68456414-139e-4c9b-928c-1eba44e74d7f", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1573053904-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f881ca8a45234848aba4416ff1d582f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96148207-1e", "ovs_interfaceid": "96148207-1e82-4a26-baa5-65f88f889f97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.816574] env[62569]: DEBUG oslo_concurrency.lockutils [None req-46c14aef-de42-4146-88f5-06424e6e802f tempest-VolumesAdminNegativeTest-504933563 tempest-VolumesAdminNegativeTest-504933563-project-member] Lock "cefbca2e-609d-4954-bec6-52ffe095446f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.118s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.854726] env[62569]: DEBUG nova.network.neutron [req-9a4ed449-afe2-421a-9b74-d5d94d3dfb60 req-cbf68f42-6706-4c87-827b-ae789d541497 service nova] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Updated VIF entry in instance network info cache for port 3d532085-b731-4296-adce-01e80852415b. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 842.855131] env[62569]: DEBUG nova.network.neutron [req-9a4ed449-afe2-421a-9b74-d5d94d3dfb60 req-cbf68f42-6706-4c87-827b-ae789d541497 service nova] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Updating instance_info_cache with network_info: [{"id": "3d532085-b731-4296-adce-01e80852415b", "address": "fa:16:3e:66:17:5b", "network": {"id": "8ea5e138-813b-4c5c-88a1-b1ea79807f10", "bridge": "br-int", "label": "tempest-ImagesTestJSON-524163445-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b20340a1ce0447cae1bfd8b7c28928f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d532085-b7", "ovs_interfaceid": "3d532085-b731-4296-adce-01e80852415b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.866518] env[62569]: DEBUG nova.compute.manager [req-74df9399-5097-496e-b85d-f2a66b1caec7 req-0f6c72e7-734f-4449-ae64-807372f28cca service nova] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Received event network-vif-plugged-030d3c5e-456b-4a2d-a229-29d169db3996 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 842.866518] env[62569]: DEBUG oslo_concurrency.lockutils [req-74df9399-5097-496e-b85d-f2a66b1caec7 req-0f6c72e7-734f-4449-ae64-807372f28cca service nova] Acquiring lock "c8edf395-ea21-4ba8-991e-0f43fe4ee830-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.866736] env[62569]: DEBUG oslo_concurrency.lockutils [req-74df9399-5097-496e-b85d-f2a66b1caec7 req-0f6c72e7-734f-4449-ae64-807372f28cca service nova] Lock "c8edf395-ea21-4ba8-991e-0f43fe4ee830-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.866919] env[62569]: DEBUG oslo_concurrency.lockutils [req-74df9399-5097-496e-b85d-f2a66b1caec7 req-0f6c72e7-734f-4449-ae64-807372f28cca service nova] Lock "c8edf395-ea21-4ba8-991e-0f43fe4ee830-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.867088] env[62569]: DEBUG nova.compute.manager [req-74df9399-5097-496e-b85d-f2a66b1caec7 req-0f6c72e7-734f-4449-ae64-807372f28cca service nova] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] No waiting events found dispatching network-vif-plugged-030d3c5e-456b-4a2d-a229-29d169db3996 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 842.867265] env[62569]: WARNING nova.compute.manager [req-74df9399-5097-496e-b85d-f2a66b1caec7 req-0f6c72e7-734f-4449-ae64-807372f28cca service nova] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Received unexpected event network-vif-plugged-030d3c5e-456b-4a2d-a229-29d169db3996 for instance with vm_state building and task_state spawning. [ 842.893288] env[62569]: DEBUG nova.network.neutron [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Successfully updated port: 030d3c5e-456b-4a2d-a229-29d169db3996 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 842.999380] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52844c82-d367-1d31-8e22-61083e68d1e7, 'name': SearchDatastore_Task, 'duration_secs': 0.022161} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.999380] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.999380] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 842.999756] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.000035] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.000522] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 843.004051] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0213d8af-f044-45d2-adf6-d551d24718de {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.015432] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 843.015893] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 843.016576] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1a81b2b-ab55-485c-8999-1b9ef87dbba5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.024308] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 843.024308] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52971ba3-eaa8-440f-5c5b-fdd98241671b" [ 843.024308] env[62569]: _type = "Task" [ 843.024308] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.034036] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52971ba3-eaa8-440f-5c5b-fdd98241671b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.081088] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2389370-52ee-4fbc-9af6-cb27d6b17fa5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.090098] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40983c4f-b8f8-4856-9360-1e0a0110b6fd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.126389] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f14b119-4baf-4da6-945f-de4ddaa8b483 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.136831] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d2c4c74-872b-4c56-8baf-ff098a8be28d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.152527] env[62569]: DEBUG nova.compute.provider_tree [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 843.173115] env[62569]: DEBUG oslo_concurrency.lockutils [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Releasing lock "refresh_cache-a5b3c345-149b-4dc3-84ff-fe72183efadb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.358583] env[62569]: DEBUG oslo_concurrency.lockutils [req-9a4ed449-afe2-421a-9b74-d5d94d3dfb60 req-cbf68f42-6706-4c87-827b-ae789d541497 service nova] Releasing lock "refresh_cache-05db90a6-3faf-4878-b782-ed17df47ed5f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.399909] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "refresh_cache-c8edf395-ea21-4ba8-991e-0f43fe4ee830" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.399909] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired lock "refresh_cache-c8edf395-ea21-4ba8-991e-0f43fe4ee830" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.399909] env[62569]: DEBUG nova.network.neutron [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 843.536289] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52971ba3-eaa8-440f-5c5b-fdd98241671b, 'name': SearchDatastore_Task, 'duration_secs': 0.012767} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.537170] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d95c2fc-65c1-41c2-b77e-4058faf16955 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.544907] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 843.544907] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d56f03-8f50-a0d0-fda5-d35c6194d907" [ 843.544907] env[62569]: _type = "Task" [ 843.544907] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.554485] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d56f03-8f50-a0d0-fda5-d35c6194d907, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.657077] env[62569]: DEBUG nova.scheduler.client.report [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 843.970853] env[62569]: DEBUG nova.network.neutron [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.057682] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d56f03-8f50-a0d0-fda5-d35c6194d907, 'name': SearchDatastore_Task, 'duration_secs': 0.014796} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.060277] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.060277] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 05db90a6-3faf-4878-b782-ed17df47ed5f/05db90a6-3faf-4878-b782-ed17df47ed5f.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 844.060277] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-243c8639-d6a3-49ae-948b-3a8f8fd80008 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.069232] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 844.069232] env[62569]: value = "task-1250178" [ 844.069232] env[62569]: _type = "Task" [ 844.069232] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.079976] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250178, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.162679] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.892s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.164874] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.068s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.165119] env[62569]: DEBUG nova.objects.instance [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lazy-loading 'resources' on Instance uuid eb70341b-4282-4eca-b6a2-374db7c521c5 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 844.203060] env[62569]: INFO nova.scheduler.client.report [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Deleted allocations for instance 77a1b192-6aff-4fee-93d7-57cebcdce626 [ 844.234651] env[62569]: DEBUG nova.network.neutron [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Updating instance_info_cache with network_info: [{"id": "030d3c5e-456b-4a2d-a229-29d169db3996", "address": "fa:16:3e:7b:7b:17", "network": {"id": "30d3bf4a-8f55-492f-a049-181e29731818", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1244949475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bdba6022e3d4697a336ca28ca4eccec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap030d3c5e-45", "ovs_interfaceid": "030d3c5e-456b-4a2d-a229-29d169db3996", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.362569] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6cb396c6-b71b-4fca-89f0-7c9087a125ba tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "a0b406c3-9466-41bd-9de1-e675cab2ceef" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.362846] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6cb396c6-b71b-4fca-89f0-7c9087a125ba tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "a0b406c3-9466-41bd-9de1-e675cab2ceef" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.584531] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250178, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.715911] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 844.717509] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-21b50f05-bf8e-420a-a552-34ba7595dc47 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.719850] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f658f1c-60a3-49fd-ad96-047abdf8e501 tempest-ServersTestFqdnHostnames-499688903 tempest-ServersTestFqdnHostnames-499688903-project-member] Lock "77a1b192-6aff-4fee-93d7-57cebcdce626" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.255s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.731571] env[62569]: DEBUG oslo_vmware.api [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Waiting for the task: (returnval){ [ 844.731571] env[62569]: value = "task-1250179" [ 844.731571] env[62569]: _type = "Task" [ 844.731571] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.737268] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Releasing lock "refresh_cache-c8edf395-ea21-4ba8-991e-0f43fe4ee830" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.737382] env[62569]: DEBUG nova.compute.manager [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Instance network_info: |[{"id": "030d3c5e-456b-4a2d-a229-29d169db3996", "address": "fa:16:3e:7b:7b:17", "network": {"id": "30d3bf4a-8f55-492f-a049-181e29731818", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1244949475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bdba6022e3d4697a336ca28ca4eccec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap030d3c5e-45", "ovs_interfaceid": "030d3c5e-456b-4a2d-a229-29d169db3996", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 844.744556] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7b:7b:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '983826cf-6390-4ec6-bf97-30a1060947fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '030d3c5e-456b-4a2d-a229-29d169db3996', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 844.757385] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Creating folder: Project (8bdba6022e3d4697a336ca28ca4eccec). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 844.758237] env[62569]: DEBUG oslo_vmware.api [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250179, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.762768] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0f9d98de-5ae5-46d8-9923-d95262a8ab59 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.775846] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Created folder: Project (8bdba6022e3d4697a336ca28ca4eccec) in parent group-v269330. [ 844.776727] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Creating folder: Instances. Parent ref: group-v269434. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 844.776727] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1adc047e-88cc-4384-b5a7-8502133e4b1b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.789060] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Created folder: Instances in parent group-v269434. [ 844.789402] env[62569]: DEBUG oslo.service.loopingcall [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.789696] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 844.790349] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f2d78946-1536-4bd2-97e8-bf994fc7a9ef {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.816308] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 844.816308] env[62569]: value = "task-1250182" [ 844.816308] env[62569]: _type = "Task" [ 844.816308] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.827854] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250182, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.866695] env[62569]: INFO nova.compute.manager [None req-6cb396c6-b71b-4fca-89f0-7c9087a125ba tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Detaching volume 025e7ad1-7cda-4045-99fb-506305a5e0cf [ 844.917959] env[62569]: DEBUG nova.compute.manager [req-b434d486-3fff-45fd-a45d-8c0ef3d0f3cf req-fe6dab20-a60e-4241-94b2-5da408114855 service nova] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Received event network-changed-030d3c5e-456b-4a2d-a229-29d169db3996 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 844.918153] env[62569]: DEBUG nova.compute.manager [req-b434d486-3fff-45fd-a45d-8c0ef3d0f3cf req-fe6dab20-a60e-4241-94b2-5da408114855 service nova] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Refreshing instance network info cache due to event network-changed-030d3c5e-456b-4a2d-a229-29d169db3996. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 844.920131] env[62569]: DEBUG oslo_concurrency.lockutils [req-b434d486-3fff-45fd-a45d-8c0ef3d0f3cf req-fe6dab20-a60e-4241-94b2-5da408114855 service nova] Acquiring lock "refresh_cache-c8edf395-ea21-4ba8-991e-0f43fe4ee830" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.920131] env[62569]: DEBUG oslo_concurrency.lockutils [req-b434d486-3fff-45fd-a45d-8c0ef3d0f3cf req-fe6dab20-a60e-4241-94b2-5da408114855 service nova] Acquired lock "refresh_cache-c8edf395-ea21-4ba8-991e-0f43fe4ee830" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.920730] env[62569]: DEBUG nova.network.neutron [req-b434d486-3fff-45fd-a45d-8c0ef3d0f3cf req-fe6dab20-a60e-4241-94b2-5da408114855 service nova] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Refreshing network info cache for port 030d3c5e-456b-4a2d-a229-29d169db3996 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 844.941264] env[62569]: INFO nova.virt.block_device [None req-6cb396c6-b71b-4fca-89f0-7c9087a125ba tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Attempting to driver detach volume 025e7ad1-7cda-4045-99fb-506305a5e0cf from mountpoint /dev/sdb [ 844.941747] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cb396c6-b71b-4fca-89f0-7c9087a125ba tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Volume detach. Driver type: vmdk {{(pid=62569) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 844.941893] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cb396c6-b71b-4fca-89f0-7c9087a125ba tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269399', 'volume_id': '025e7ad1-7cda-4045-99fb-506305a5e0cf', 'name': 'volume-025e7ad1-7cda-4045-99fb-506305a5e0cf', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a0b406c3-9466-41bd-9de1-e675cab2ceef', 'attached_at': '', 'detached_at': '', 'volume_id': '025e7ad1-7cda-4045-99fb-506305a5e0cf', 'serial': '025e7ad1-7cda-4045-99fb-506305a5e0cf'} {{(pid=62569) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 844.943395] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce26ff30-a6a0-4ab2-9c68-a1b596290c39 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.980680] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2b42459-f40d-4e44-bab9-58f6a2cf4cf8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.988160] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a449df96-91a4-4e0b-b38e-b31497cb76c3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.017295] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-250f712b-3c85-4095-8758-baad25a2e8ed {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.035376] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cb396c6-b71b-4fca-89f0-7c9087a125ba tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] The volume has not been displaced from its original location: [datastore1] volume-025e7ad1-7cda-4045-99fb-506305a5e0cf/volume-025e7ad1-7cda-4045-99fb-506305a5e0cf.vmdk. No consolidation needed. {{(pid=62569) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 845.040952] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cb396c6-b71b-4fca-89f0-7c9087a125ba tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Reconfiguring VM instance instance-00000034 to detach disk 2001 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 845.043965] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-343a9e46-83f2-4ccb-b998-952dc739f94f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.066672] env[62569]: DEBUG oslo_vmware.api [None req-6cb396c6-b71b-4fca-89f0-7c9087a125ba tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 845.066672] env[62569]: value = "task-1250183" [ 845.066672] env[62569]: _type = "Task" [ 845.066672] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.072052] env[62569]: DEBUG oslo_vmware.api [None req-6cb396c6-b71b-4fca-89f0-7c9087a125ba tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250183, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.083702] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250178, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.600551} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.083953] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 05db90a6-3faf-4878-b782-ed17df47ed5f/05db90a6-3faf-4878-b782-ed17df47ed5f.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 845.084170] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 845.084461] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4b95525f-de30-434a-a7d6-bbec23eb6b0a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.090886] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 845.090886] env[62569]: value = "task-1250184" [ 845.090886] env[62569]: _type = "Task" [ 845.090886] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.099565] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250184, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.127173] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc27b9d7-f996-45db-a12b-e7548ea5aab0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.137299] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a36a47d1-0fc0-4a29-84f7-803217115de1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.166162] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18aef231-80cc-4b69-a40c-23b0687db712 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.173869] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e5ca1b4-64d6-4741-8ad6-39ce6afb49eb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.189655] env[62569]: DEBUG nova.compute.provider_tree [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 845.241093] env[62569]: DEBUG oslo_vmware.api [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250179, 'name': PowerOffVM_Task, 'duration_secs': 0.315973} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.241607] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 845.242413] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91103186-ff1f-4ab7-bf2b-ea302f5ad8df {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.261564] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-986c510f-4106-4f0f-8c74-55b3d20b7531 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.294374] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 845.294655] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cc371ed5-60d6-4bce-9d5e-93f830c1c39d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.300366] env[62569]: DEBUG oslo_vmware.api [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Waiting for the task: (returnval){ [ 845.300366] env[62569]: value = "task-1250185" [ 845.300366] env[62569]: _type = "Task" [ 845.300366] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.308024] env[62569]: DEBUG oslo_vmware.api [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250185, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.328016] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250182, 'name': CreateVM_Task, 'duration_secs': 0.412976} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.328016] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 845.328016] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.328016] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.328016] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 845.328016] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b53db54d-a3c0-435d-8075-c43860c7778a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.333750] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 845.333750] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52c21154-4748-fe6a-ebfe-d3d8cfa9951a" [ 845.333750] env[62569]: _type = "Task" [ 845.333750] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.340981] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52c21154-4748-fe6a-ebfe-d3d8cfa9951a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.576610] env[62569]: DEBUG oslo_vmware.api [None req-6cb396c6-b71b-4fca-89f0-7c9087a125ba tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250183, 'name': ReconfigVM_Task, 'duration_secs': 0.242611} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.577433] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cb396c6-b71b-4fca-89f0-7c9087a125ba tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Reconfigured VM instance instance-00000034 to detach disk 2001 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 845.583077] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ddfd6146-e639-43fa-98f6-900f4b6144d0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.608751] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250184, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083834} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.610316] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 845.610843] env[62569]: DEBUG oslo_vmware.api [None req-6cb396c6-b71b-4fca-89f0-7c9087a125ba tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 845.610843] env[62569]: value = "task-1250186" [ 845.610843] env[62569]: _type = "Task" [ 845.610843] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.611765] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d567b73-5043-4e15-9122-e3ee73968a50 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.636852] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 05db90a6-3faf-4878-b782-ed17df47ed5f/05db90a6-3faf-4878-b782-ed17df47ed5f.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 845.640763] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1181df34-42f6-4663-a137-b504fc80ecf4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.658183] env[62569]: DEBUG oslo_vmware.api [None req-6cb396c6-b71b-4fca-89f0-7c9087a125ba tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250186, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.663352] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 845.663352] env[62569]: value = "task-1250187" [ 845.663352] env[62569]: _type = "Task" [ 845.663352] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.674061] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250187, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.694120] env[62569]: DEBUG nova.scheduler.client.report [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 845.772163] env[62569]: DEBUG nova.network.neutron [req-b434d486-3fff-45fd-a45d-8c0ef3d0f3cf req-fe6dab20-a60e-4241-94b2-5da408114855 service nova] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Updated VIF entry in instance network info cache for port 030d3c5e-456b-4a2d-a229-29d169db3996. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 845.773335] env[62569]: DEBUG nova.network.neutron [req-b434d486-3fff-45fd-a45d-8c0ef3d0f3cf req-fe6dab20-a60e-4241-94b2-5da408114855 service nova] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Updating instance_info_cache with network_info: [{"id": "030d3c5e-456b-4a2d-a229-29d169db3996", "address": "fa:16:3e:7b:7b:17", "network": {"id": "30d3bf4a-8f55-492f-a049-181e29731818", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1244949475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bdba6022e3d4697a336ca28ca4eccec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap030d3c5e-45", "ovs_interfaceid": "030d3c5e-456b-4a2d-a229-29d169db3996", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.817634] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] VM already powered off {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 845.817922] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 845.818231] env[62569]: DEBUG oslo_concurrency.lockutils [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.818413] env[62569]: DEBUG oslo_concurrency.lockutils [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.818620] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 845.819354] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-345bb23b-c136-4bff-9ff9-ac286f7b127a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.828324] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 845.828456] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 845.829330] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b650e99-59a5-4831-91bb-a0c896bb3b8b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.836037] env[62569]: DEBUG oslo_vmware.api [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Waiting for the task: (returnval){ [ 845.836037] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f0024f-f281-fb5b-368b-64caf6aecfcf" [ 845.836037] env[62569]: _type = "Task" [ 845.836037] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.852031] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52c21154-4748-fe6a-ebfe-d3d8cfa9951a, 'name': SearchDatastore_Task, 'duration_secs': 0.01718} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.855640] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.855901] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 845.856177] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.856329] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.856514] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 845.856798] env[62569]: DEBUG oslo_vmware.api [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f0024f-f281-fb5b-368b-64caf6aecfcf, 'name': SearchDatastore_Task, 'duration_secs': 0.008356} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.857039] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-670d371e-6d75-4d5e-9610-a742aabc1097 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.861156] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62f54178-197d-4bff-a8b1-083f796e05ab {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.867070] env[62569]: DEBUG oslo_vmware.api [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Waiting for the task: (returnval){ [ 845.867070] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]521ec6df-552c-54a3-4743-a9ab66553f24" [ 845.867070] env[62569]: _type = "Task" [ 845.867070] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.872209] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 845.872423] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 845.873600] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a1fe16b-78e0-4019-9375-ec5e8495bdc6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.880072] env[62569]: DEBUG oslo_vmware.api [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]521ec6df-552c-54a3-4743-a9ab66553f24, 'name': SearchDatastore_Task} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.880847] env[62569]: DEBUG oslo_concurrency.lockutils [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.881233] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] a5b3c345-149b-4dc3-84ff-fe72183efadb/f77800cf-af0d-4e9c-b312-2c59488f2c7c-rescue.vmdk. {{(pid=62569) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 845.881568] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-72563cde-dfd6-4bb4-90d9-9d6ed9ebe4a0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.888098] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 845.888098] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52984349-e4e3-877c-b163-ad2f06c43389" [ 845.888098] env[62569]: _type = "Task" [ 845.888098] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.890603] env[62569]: DEBUG oslo_vmware.api [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Waiting for the task: (returnval){ [ 845.890603] env[62569]: value = "task-1250188" [ 845.890603] env[62569]: _type = "Task" [ 845.890603] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.898205] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52984349-e4e3-877c-b163-ad2f06c43389, 'name': SearchDatastore_Task, 'duration_secs': 0.008028} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.899541] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d70186b0-9033-4d42-8f1c-3f5fc3f2425c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.905298] env[62569]: DEBUG oslo_vmware.api [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250188, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.908625] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 845.908625] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]522c198f-2780-bd69-ecee-562c0f7c31b9" [ 845.908625] env[62569]: _type = "Task" [ 845.908625] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.917645] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]522c198f-2780-bd69-ecee-562c0f7c31b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.127244] env[62569]: DEBUG oslo_vmware.api [None req-6cb396c6-b71b-4fca-89f0-7c9087a125ba tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250186, 'name': ReconfigVM_Task, 'duration_secs': 0.148645} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.127888] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cb396c6-b71b-4fca-89f0-7c9087a125ba tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269399', 'volume_id': '025e7ad1-7cda-4045-99fb-506305a5e0cf', 'name': 'volume-025e7ad1-7cda-4045-99fb-506305a5e0cf', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a0b406c3-9466-41bd-9de1-e675cab2ceef', 'attached_at': '', 'detached_at': '', 'volume_id': '025e7ad1-7cda-4045-99fb-506305a5e0cf', 'serial': '025e7ad1-7cda-4045-99fb-506305a5e0cf'} {{(pid=62569) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 846.179027] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250187, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.202902] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.036s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.205674] env[62569]: DEBUG oslo_concurrency.lockutils [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.688s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.208936] env[62569]: INFO nova.compute.claims [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 846.237296] env[62569]: INFO nova.scheduler.client.report [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Deleted allocations for instance eb70341b-4282-4eca-b6a2-374db7c521c5 [ 846.277821] env[62569]: DEBUG oslo_concurrency.lockutils [req-b434d486-3fff-45fd-a45d-8c0ef3d0f3cf req-fe6dab20-a60e-4241-94b2-5da408114855 service nova] Releasing lock "refresh_cache-c8edf395-ea21-4ba8-991e-0f43fe4ee830" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.403427] env[62569]: DEBUG oslo_vmware.api [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250188, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.44488} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.403717] env[62569]: INFO nova.virt.vmwareapi.ds_util [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] a5b3c345-149b-4dc3-84ff-fe72183efadb/f77800cf-af0d-4e9c-b312-2c59488f2c7c-rescue.vmdk. [ 846.404586] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b937a12-c63f-498c-80aa-802959f1118b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.440200] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] a5b3c345-149b-4dc3-84ff-fe72183efadb/f77800cf-af0d-4e9c-b312-2c59488f2c7c-rescue.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.441724] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed7311d0-37bb-4926-95f9-e8a129e57490 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.464944] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]522c198f-2780-bd69-ecee-562c0f7c31b9, 'name': SearchDatastore_Task, 'duration_secs': 0.009219} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.465792] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.466280] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] c8edf395-ea21-4ba8-991e-0f43fe4ee830/c8edf395-ea21-4ba8-991e-0f43fe4ee830.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 846.466632] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e14d1cd1-e0e3-4731-98fe-80de3c6d01c8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.470364] env[62569]: DEBUG oslo_vmware.api [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Waiting for the task: (returnval){ [ 846.470364] env[62569]: value = "task-1250189" [ 846.470364] env[62569]: _type = "Task" [ 846.470364] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.475253] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 846.475253] env[62569]: value = "task-1250190" [ 846.475253] env[62569]: _type = "Task" [ 846.475253] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.482226] env[62569]: DEBUG oslo_vmware.api [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250189, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.487209] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250190, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.676054] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250187, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.698832] env[62569]: DEBUG nova.objects.instance [None req-6cb396c6-b71b-4fca-89f0-7c9087a125ba tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lazy-loading 'flavor' on Instance uuid a0b406c3-9466-41bd-9de1-e675cab2ceef {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 846.751681] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4b5176cc-2107-4eea-971f-59e6276d9bc5 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "eb70341b-4282-4eca-b6a2-374db7c521c5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.192s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.985601] env[62569]: DEBUG oslo_vmware.api [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250189, 'name': ReconfigVM_Task, 'duration_secs': 0.41937} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.989047] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Reconfigured VM instance instance-0000004c to attach disk [datastore2] a5b3c345-149b-4dc3-84ff-fe72183efadb/f77800cf-af0d-4e9c-b312-2c59488f2c7c-rescue.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 846.989712] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250190, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.468596} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.990410] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddd3f764-c613-416e-8250-ed3f1587cf9a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.993055] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] c8edf395-ea21-4ba8-991e-0f43fe4ee830/c8edf395-ea21-4ba8-991e-0f43fe4ee830.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 846.993292] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 846.993557] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1d4e42c5-8acc-4117-978d-40b292c58232 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.021664] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a122ad1d-06a6-4251-b68b-e7ff07608517 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.034274] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 847.034274] env[62569]: value = "task-1250191" [ 847.034274] env[62569]: _type = "Task" [ 847.034274] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.040896] env[62569]: DEBUG oslo_vmware.api [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Waiting for the task: (returnval){ [ 847.040896] env[62569]: value = "task-1250192" [ 847.040896] env[62569]: _type = "Task" [ 847.040896] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.044867] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250191, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.054498] env[62569]: DEBUG oslo_vmware.api [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250192, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.174074] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250187, 'name': ReconfigVM_Task, 'duration_secs': 1.266755} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.174458] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 05db90a6-3faf-4878-b782-ed17df47ed5f/05db90a6-3faf-4878-b782-ed17df47ed5f.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.175056] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-99d975b7-fa19-4730-a7a2-195a50b7f500 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.181620] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 847.181620] env[62569]: value = "task-1250193" [ 847.181620] env[62569]: _type = "Task" [ 847.181620] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.189860] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250193, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.534846] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b02c80c-c840-470e-8160-1b69d55483b7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.553951] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250191, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07038} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.556105] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 847.556105] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2582ca80-5940-4d9c-aeb4-14262b36e461 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.563224] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-686f8f03-0605-4cf2-b7f3-1e7d542853ef {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.567327] env[62569]: DEBUG oslo_vmware.api [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250192, 'name': ReconfigVM_Task, 'duration_secs': 0.230123} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.568609] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 847.568609] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-245576d6-2693-49c4-837b-2b9a95dedd41 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.605910] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef58b352-a389-4293-9b91-103e9ea9e4f7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.618028] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] c8edf395-ea21-4ba8-991e-0f43fe4ee830/c8edf395-ea21-4ba8-991e-0f43fe4ee830.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 847.619477] env[62569]: DEBUG oslo_concurrency.lockutils [None req-611c62e3-2421-478c-8dce-ddb34e795534 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "a0b406c3-9466-41bd-9de1-e675cab2ceef" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.619830] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-899ed730-6267-427f-828a-c3efa92ea624 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.640068] env[62569]: DEBUG oslo_vmware.api [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Waiting for the task: (returnval){ [ 847.640068] env[62569]: value = "task-1250194" [ 847.640068] env[62569]: _type = "Task" [ 847.640068] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.650408] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19e5d3d8-fa9f-45cd-89a8-d751cf693872 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.654796] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 847.654796] env[62569]: value = "task-1250195" [ 847.654796] env[62569]: _type = "Task" [ 847.654796] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.658484] env[62569]: DEBUG oslo_vmware.api [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250194, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.670846] env[62569]: DEBUG nova.compute.provider_tree [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 847.678307] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250195, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.693199] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250193, 'name': Rename_Task, 'duration_secs': 0.170196} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.693199] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 847.693199] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0223e2ab-4208-455d-a10b-95890e053c29 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.700521] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 847.700521] env[62569]: value = "task-1250196" [ 847.700521] env[62569]: _type = "Task" [ 847.700521] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.709225] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6cb396c6-b71b-4fca-89f0-7c9087a125ba tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "a0b406c3-9466-41bd-9de1-e675cab2ceef" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.346s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.711920] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250196, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.712128] env[62569]: DEBUG oslo_concurrency.lockutils [None req-611c62e3-2421-478c-8dce-ddb34e795534 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "a0b406c3-9466-41bd-9de1-e675cab2ceef" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.093s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.712406] env[62569]: DEBUG nova.compute.manager [None req-611c62e3-2421-478c-8dce-ddb34e795534 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 847.713513] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea7a23e3-c08e-45eb-b1d6-176a5216973e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.719917] env[62569]: DEBUG nova.compute.manager [None req-611c62e3-2421-478c-8dce-ddb34e795534 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62569) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 847.720542] env[62569]: DEBUG nova.objects.instance [None req-611c62e3-2421-478c-8dce-ddb34e795534 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lazy-loading 'flavor' on Instance uuid a0b406c3-9466-41bd-9de1-e675cab2ceef {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 848.159096] env[62569]: DEBUG oslo_vmware.api [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250194, 'name': PowerOnVM_Task, 'duration_secs': 0.39527} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.159096] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 848.168697] env[62569]: DEBUG nova.compute.manager [None req-916edd11-0ed0-4aad-96b9-46f69c1709da tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 848.170455] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a104bd8c-a7cd-4ab9-80b5-cc4703bb083a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.188972] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250195, 'name': ReconfigVM_Task, 'duration_secs': 0.421956} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.188972] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Reconfigured VM instance instance-0000004e to attach disk [datastore1] c8edf395-ea21-4ba8-991e-0f43fe4ee830/c8edf395-ea21-4ba8-991e-0f43fe4ee830.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 848.190098] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f737b9a1-3e07-4ec9-abd9-1e294e342f60 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.200188] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 848.200188] env[62569]: value = "task-1250197" [ 848.200188] env[62569]: _type = "Task" [ 848.200188] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.208893] env[62569]: ERROR nova.scheduler.client.report [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [req-a8a8d4ca-b20f-4f89-bb93-51d9be044ff8] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID fa06556a-5785-4014-b8bd-bc240a0cf716. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a8a8d4ca-b20f-4f89-bb93-51d9be044ff8"}]} [ 848.218580] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250197, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.226810] env[62569]: DEBUG oslo_vmware.api [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250196, 'name': PowerOnVM_Task, 'duration_secs': 0.516465} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.227129] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 848.227356] env[62569]: INFO nova.compute.manager [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Took 9.37 seconds to spawn the instance on the hypervisor. [ 848.228461] env[62569]: DEBUG nova.compute.manager [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 848.228461] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c01e40be-015d-49eb-938d-287100e19858 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.240265] env[62569]: DEBUG nova.scheduler.client.report [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Refreshing inventories for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 848.266161] env[62569]: DEBUG nova.scheduler.client.report [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Updating ProviderTree inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 848.266584] env[62569]: DEBUG nova.compute.provider_tree [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 848.291664] env[62569]: DEBUG nova.scheduler.client.report [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Refreshing aggregate associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, aggregates: None {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 848.319324] env[62569]: DEBUG nova.scheduler.client.report [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Refreshing trait associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 848.582400] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed95b0fd-01b8-42f6-a2b7-d65dd52a5472 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.597090] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f35d43a4-560d-4c16-aa2c-338004b99d01 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.638459] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51e73580-ab23-4284-9472-eb404b9cebdf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.646166] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7987c86-ddd7-47a3-8de3-97360c1865e7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.661772] env[62569]: DEBUG nova.compute.provider_tree [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 848.710126] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250197, 'name': Rename_Task, 'duration_secs': 0.305321} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.710400] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 848.710653] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c5fe683-41af-4a94-bfb5-0e9ee4dfc0c7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.717505] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 848.717505] env[62569]: value = "task-1250198" [ 848.717505] env[62569]: _type = "Task" [ 848.717505] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.726580] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250198, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.735652] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-611c62e3-2421-478c-8dce-ddb34e795534 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 848.735718] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-007cc4b9-bef1-4eca-b67f-a468ec95c066 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.742791] env[62569]: DEBUG oslo_vmware.api [None req-611c62e3-2421-478c-8dce-ddb34e795534 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 848.742791] env[62569]: value = "task-1250199" [ 848.742791] env[62569]: _type = "Task" [ 848.742791] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.757434] env[62569]: DEBUG oslo_vmware.api [None req-611c62e3-2421-478c-8dce-ddb34e795534 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250199, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.757775] env[62569]: INFO nova.compute.manager [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Took 38.47 seconds to build instance. [ 849.199046] env[62569]: DEBUG nova.scheduler.client.report [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Updated inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with generation 88 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 849.200629] env[62569]: DEBUG nova.compute.provider_tree [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Updating resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 generation from 88 to 89 during operation: update_inventory {{(pid=62569) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 849.200629] env[62569]: DEBUG nova.compute.provider_tree [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 849.231964] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250198, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.255375] env[62569]: DEBUG oslo_vmware.api [None req-611c62e3-2421-478c-8dce-ddb34e795534 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250199, 'name': PowerOffVM_Task, 'duration_secs': 0.235994} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.255375] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-611c62e3-2421-478c-8dce-ddb34e795534 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 849.255375] env[62569]: DEBUG nova.compute.manager [None req-611c62e3-2421-478c-8dce-ddb34e795534 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 849.255583] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a00c2086-4afa-4580-9a5d-a90d8ff82d72 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.260080] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32737218-acca-4b55-8206-0caad8e06319 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "05db90a6-3faf-4878-b782-ed17df47ed5f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.390s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.650448] env[62569]: DEBUG oslo_vmware.rw_handles [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bb5346-cfae-5dda-f4b1-7af62ff83a9f/disk-0.vmdk. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 849.650690] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-955be975-95ef-4dfc-8437-488468a458b2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.660236] env[62569]: DEBUG oslo_vmware.rw_handles [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bb5346-cfae-5dda-f4b1-7af62ff83a9f/disk-0.vmdk is in state: ready. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 849.660236] env[62569]: ERROR oslo_vmware.rw_handles [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bb5346-cfae-5dda-f4b1-7af62ff83a9f/disk-0.vmdk due to incomplete transfer. [ 849.660236] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-8c2c1f28-c030-4dac-8723-4805818a1034 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.665579] env[62569]: DEBUG oslo_vmware.rw_handles [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bb5346-cfae-5dda-f4b1-7af62ff83a9f/disk-0.vmdk. {{(pid=62569) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 849.670174] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Uploaded image 45f3d90c-b69b-4905-993e-439633bfeb29 to the Glance image server {{(pid=62569) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 849.670174] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Destroying the VM {{(pid=62569) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 849.670174] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-d539a9d8-b298-441e-a25e-bb5200225def {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.677340] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 849.677340] env[62569]: value = "task-1250200" [ 849.677340] env[62569]: _type = "Task" [ 849.677340] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.685998] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250200, 'name': Destroy_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.709031] env[62569]: DEBUG oslo_concurrency.lockutils [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.503s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.709031] env[62569]: DEBUG nova.compute.manager [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 849.711498] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.077s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.711981] env[62569]: DEBUG nova.objects.instance [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Lazy-loading 'resources' on Instance uuid 390bcf25-689d-46ad-bffb-3670c3729397 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 849.728446] env[62569]: DEBUG oslo_vmware.api [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250198, 'name': PowerOnVM_Task, 'duration_secs': 0.723588} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.729031] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 849.729031] env[62569]: INFO nova.compute.manager [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Took 7.42 seconds to spawn the instance on the hypervisor. [ 849.729348] env[62569]: DEBUG nova.compute.manager [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 849.729916] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afeb2a5d-3e27-4f47-ab23-335788bb8c0a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.771310] env[62569]: DEBUG oslo_concurrency.lockutils [None req-611c62e3-2421-478c-8dce-ddb34e795534 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "a0b406c3-9466-41bd-9de1-e675cab2ceef" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.059s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.142730] env[62569]: DEBUG nova.compute.manager [req-159629b8-3cfe-4e23-8611-71477485f265 req-d76139ea-4b51-48aa-add1-5afa3509099c service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Received event network-changed-96148207-1e82-4a26-baa5-65f88f889f97 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 850.142940] env[62569]: DEBUG nova.compute.manager [req-159629b8-3cfe-4e23-8611-71477485f265 req-d76139ea-4b51-48aa-add1-5afa3509099c service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Refreshing instance network info cache due to event network-changed-96148207-1e82-4a26-baa5-65f88f889f97. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 850.143195] env[62569]: DEBUG oslo_concurrency.lockutils [req-159629b8-3cfe-4e23-8611-71477485f265 req-d76139ea-4b51-48aa-add1-5afa3509099c service nova] Acquiring lock "refresh_cache-a5b3c345-149b-4dc3-84ff-fe72183efadb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.143367] env[62569]: DEBUG oslo_concurrency.lockutils [req-159629b8-3cfe-4e23-8611-71477485f265 req-d76139ea-4b51-48aa-add1-5afa3509099c service nova] Acquired lock "refresh_cache-a5b3c345-149b-4dc3-84ff-fe72183efadb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.143534] env[62569]: DEBUG nova.network.neutron [req-159629b8-3cfe-4e23-8611-71477485f265 req-d76139ea-4b51-48aa-add1-5afa3509099c service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Refreshing network info cache for port 96148207-1e82-4a26-baa5-65f88f889f97 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 850.188132] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250200, 'name': Destroy_Task} progress is 33%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.214956] env[62569]: DEBUG nova.compute.utils [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 850.220119] env[62569]: DEBUG nova.compute.manager [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 850.220359] env[62569]: DEBUG nova.network.neutron [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 850.251709] env[62569]: INFO nova.compute.manager [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Took 36.74 seconds to build instance. [ 850.287675] env[62569]: DEBUG nova.policy [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1418d7ccc2a4324aeb586f89c2f817b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4ba7e96d226942bd99ab29fe703c6fea', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 850.500018] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab9b8d20-510f-4be0-aa08-fcbb0f814274 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.508770] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0e27b3a-22da-4664-8c62-7d1323c6be1c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.551252] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5178a04-f722-4bd3-8f04-5502f4246274 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.572806] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8146a2-56c5-4544-b668-ad110a86cf4e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.572806] env[62569]: DEBUG nova.network.neutron [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Successfully created port: 8e55dca6-f777-4333-9b49-feb5527d9a5e {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 850.588031] env[62569]: DEBUG nova.compute.provider_tree [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.693866] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250200, 'name': Destroy_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.724638] env[62569]: DEBUG nova.compute.manager [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 850.754165] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4f115f9a-da0b-4afb-bd0f-a672e5690078 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "c8edf395-ea21-4ba8-991e-0f43fe4ee830" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.242s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.073400] env[62569]: DEBUG nova.compute.manager [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 851.074380] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d3bd58a-c87f-4c78-b677-0f9851591c57 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.091167] env[62569]: DEBUG nova.scheduler.client.report [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 851.102873] env[62569]: DEBUG nova.network.neutron [req-159629b8-3cfe-4e23-8611-71477485f265 req-d76139ea-4b51-48aa-add1-5afa3509099c service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Updated VIF entry in instance network info cache for port 96148207-1e82-4a26-baa5-65f88f889f97. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 851.107044] env[62569]: DEBUG nova.network.neutron [req-159629b8-3cfe-4e23-8611-71477485f265 req-d76139ea-4b51-48aa-add1-5afa3509099c service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Updating instance_info_cache with network_info: [{"id": "96148207-1e82-4a26-baa5-65f88f889f97", "address": "fa:16:3e:03:33:1e", "network": {"id": "68456414-139e-4c9b-928c-1eba44e74d7f", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1573053904-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f881ca8a45234848aba4416ff1d582f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96148207-1e", "ovs_interfaceid": "96148207-1e82-4a26-baa5-65f88f889f97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.109770] env[62569]: DEBUG nova.objects.instance [None req-1a0d2531-4940-49b1-b1c7-b82f2d911bad tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lazy-loading 'flavor' on Instance uuid a0b406c3-9466-41bd-9de1-e675cab2ceef {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 851.192967] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250200, 'name': Destroy_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.238148] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "c8edf395-ea21-4ba8-991e-0f43fe4ee830" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.238443] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "c8edf395-ea21-4ba8-991e-0f43fe4ee830" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.238603] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "c8edf395-ea21-4ba8-991e-0f43fe4ee830-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.238831] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "c8edf395-ea21-4ba8-991e-0f43fe4ee830-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.239080] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "c8edf395-ea21-4ba8-991e-0f43fe4ee830-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.241082] env[62569]: INFO nova.compute.manager [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Terminating instance [ 851.586813] env[62569]: INFO nova.compute.manager [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] instance snapshotting [ 851.591438] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca7eeaa-87ed-4818-84ab-73596a9769c8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.596402] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.885s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.598900] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 22.021s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.599277] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.599277] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62569) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 851.599536] env[62569]: DEBUG oslo_concurrency.lockutils [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.063s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.601893] env[62569]: INFO nova.compute.claims [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 851.618240] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-350bbeda-ea74-4e10-a52a-acb2508085a3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.621284] env[62569]: DEBUG oslo_concurrency.lockutils [req-159629b8-3cfe-4e23-8611-71477485f265 req-d76139ea-4b51-48aa-add1-5afa3509099c service nova] Releasing lock "refresh_cache-a5b3c345-149b-4dc3-84ff-fe72183efadb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.624317] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca8e4b95-c680-4e90-8bf2-fd783bc27437 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.627498] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1a0d2531-4940-49b1-b1c7-b82f2d911bad tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "refresh_cache-a0b406c3-9466-41bd-9de1-e675cab2ceef" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.627692] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1a0d2531-4940-49b1-b1c7-b82f2d911bad tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquired lock "refresh_cache-a0b406c3-9466-41bd-9de1-e675cab2ceef" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.627773] env[62569]: DEBUG nova.network.neutron [None req-1a0d2531-4940-49b1-b1c7-b82f2d911bad tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 851.628098] env[62569]: DEBUG nova.objects.instance [None req-1a0d2531-4940-49b1-b1c7-b82f2d911bad tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lazy-loading 'info_cache' on Instance uuid a0b406c3-9466-41bd-9de1-e675cab2ceef {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 851.631564] env[62569]: INFO nova.scheduler.client.report [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Deleted allocations for instance 390bcf25-689d-46ad-bffb-3670c3729397 [ 851.642387] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50911c23-f2dc-42b9-a142-e81a42c37177 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.660311] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17f5f805-d8e1-4309-925c-68e36cde9002 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.669946] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f70c1467-70ee-4336-8a4e-d226ebf6d999 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.709736] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180248MB free_disk=146GB free_vcpus=48 pci_devices=None {{(pid=62569) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 851.709736] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.717705] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250200, 'name': Destroy_Task, 'duration_secs': 1.562759} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.717987] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Destroyed the VM [ 851.718490] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Deleting Snapshot of the VM instance {{(pid=62569) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 851.718490] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0278012a-d594-4d5c-a5c5-bf4ecc37c7a5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.730605] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 851.730605] env[62569]: value = "task-1250201" [ 851.730605] env[62569]: _type = "Task" [ 851.730605] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.737565] env[62569]: DEBUG nova.compute.manager [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 851.743452] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250201, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.747267] env[62569]: DEBUG nova.compute.manager [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 851.747508] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 851.748381] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da1c2d7e-d3ec-4066-863a-0bd6072afb67 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.756865] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 851.757238] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b18d64d9-075b-47bb-89c7-df900fb00e4e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.761344] env[62569]: DEBUG nova.virt.hardware [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 851.761583] env[62569]: DEBUG nova.virt.hardware [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 851.761831] env[62569]: DEBUG nova.virt.hardware [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 851.762031] env[62569]: DEBUG nova.virt.hardware [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 851.762221] env[62569]: DEBUG nova.virt.hardware [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 851.762430] env[62569]: DEBUG nova.virt.hardware [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 851.762622] env[62569]: DEBUG nova.virt.hardware [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 851.762812] env[62569]: DEBUG nova.virt.hardware [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 851.763295] env[62569]: DEBUG nova.virt.hardware [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 851.763295] env[62569]: DEBUG nova.virt.hardware [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 851.763446] env[62569]: DEBUG nova.virt.hardware [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 851.764732] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b643b1f1-be6b-4071-97a9-89f55a57bf04 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.773577] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c52caf90-b147-42a0-bf1c-aded232aac15 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.779614] env[62569]: DEBUG oslo_vmware.api [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 851.779614] env[62569]: value = "task-1250202" [ 851.779614] env[62569]: _type = "Task" [ 851.779614] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.799521] env[62569]: DEBUG oslo_vmware.api [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250202, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.133273] env[62569]: DEBUG nova.objects.base [None req-1a0d2531-4940-49b1-b1c7-b82f2d911bad tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62569) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 852.150827] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Creating Snapshot of the VM instance {{(pid=62569) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 852.151395] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-0b502a2c-e851-42b1-b121-d8135c048cee {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.153758] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8d8fcff3-a4dd-4158-a99d-e89f0080a847 tempest-ServersTestMultiNic-1556202612 tempest-ServersTestMultiNic-1556202612-project-member] Lock "390bcf25-689d-46ad-bffb-3670c3729397" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.856s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.160365] env[62569]: DEBUG oslo_vmware.api [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 852.160365] env[62569]: value = "task-1250203" [ 852.160365] env[62569]: _type = "Task" [ 852.160365] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.175970] env[62569]: DEBUG oslo_vmware.api [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250203, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.230289] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dccb5fb1-ce24-4e5e-af94-487a7aee6c09 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "interface-807d3025-d6a7-4778-a829-a61e2c7495c9-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.230289] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dccb5fb1-ce24-4e5e-af94-487a7aee6c09 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "interface-807d3025-d6a7-4778-a829-a61e2c7495c9-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.230289] env[62569]: DEBUG nova.objects.instance [None req-dccb5fb1-ce24-4e5e-af94-487a7aee6c09 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lazy-loading 'flavor' on Instance uuid 807d3025-d6a7-4778-a829-a61e2c7495c9 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 852.243637] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250201, 'name': RemoveSnapshot_Task, 'duration_secs': 0.401064} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.244576] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Deleted Snapshot of the VM instance {{(pid=62569) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 852.245348] env[62569]: DEBUG nova.compute.manager [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 852.247090] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4ec8ad9-d2f5-4e17-a83c-476173f8f162 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.293621] env[62569]: DEBUG oslo_vmware.api [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250202, 'name': PowerOffVM_Task, 'duration_secs': 0.259261} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.296886] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 852.296886] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 852.296886] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-672579f6-0ba1-41ba-9489-07853dfd559d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.305455] env[62569]: DEBUG nova.network.neutron [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Successfully updated port: 8e55dca6-f777-4333-9b49-feb5527d9a5e {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 852.363838] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 852.364089] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 852.364281] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Deleting the datastore file [datastore1] c8edf395-ea21-4ba8-991e-0f43fe4ee830 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 852.364553] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c02bd9bb-499a-4047-a6f9-be63877e111e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.371526] env[62569]: DEBUG oslo_vmware.api [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 852.371526] env[62569]: value = "task-1250205" [ 852.371526] env[62569]: _type = "Task" [ 852.371526] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.379874] env[62569]: DEBUG oslo_vmware.api [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250205, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.674972] env[62569]: DEBUG oslo_vmware.api [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250203, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.733397] env[62569]: DEBUG nova.objects.instance [None req-dccb5fb1-ce24-4e5e-af94-487a7aee6c09 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lazy-loading 'pci_requests' on Instance uuid 807d3025-d6a7-4778-a829-a61e2c7495c9 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 852.764135] env[62569]: INFO nova.compute.manager [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Shelve offloading [ 852.807670] env[62569]: DEBUG oslo_concurrency.lockutils [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "refresh_cache-31ce29fa-4f60-4404-b830-21ad196f78b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.807813] env[62569]: DEBUG oslo_concurrency.lockutils [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired lock "refresh_cache-31ce29fa-4f60-4404-b830-21ad196f78b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.807968] env[62569]: DEBUG nova.network.neutron [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 852.862250] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-813e0b73-a040-467f-8050-ed41a4b68f38 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.867068] env[62569]: DEBUG nova.network.neutron [None req-1a0d2531-4940-49b1-b1c7-b82f2d911bad tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Updating instance_info_cache with network_info: [{"id": "3da67b46-e89d-4a96-8dd9-752956b9b270", "address": "fa:16:3e:f6:2c:c2", "network": {"id": "ba6b71b1-bd7b-493e-8682-68d842fa935f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1892479706-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e9ea55b00b574a7db0530efcb498ca2b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3da67b46-e8", "ovs_interfaceid": "3da67b46-e89d-4a96-8dd9-752956b9b270", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.871978] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59953666-dbf5-46e2-855c-2aca50fcfd96 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.916665] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c31675a-80ff-4e5c-ac83-194ec82f6c4b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.920393] env[62569]: DEBUG oslo_vmware.api [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250205, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.351453} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.920906] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 852.920985] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 852.921429] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 852.921429] env[62569]: INFO nova.compute.manager [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Took 1.17 seconds to destroy the instance on the hypervisor. [ 852.921758] env[62569]: DEBUG oslo.service.loopingcall [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 852.922469] env[62569]: DEBUG nova.compute.manager [-] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 852.922580] env[62569]: DEBUG nova.network.neutron [-] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 852.929888] env[62569]: DEBUG nova.compute.manager [req-dc8563aa-87ef-4102-95c7-e9f4fee9d7f7 req-cd8d18af-b4c4-4e59-a8b4-84d5a7f4954a service nova] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Received event network-vif-plugged-8e55dca6-f777-4333-9b49-feb5527d9a5e {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 852.930183] env[62569]: DEBUG oslo_concurrency.lockutils [req-dc8563aa-87ef-4102-95c7-e9f4fee9d7f7 req-cd8d18af-b4c4-4e59-a8b4-84d5a7f4954a service nova] Acquiring lock "31ce29fa-4f60-4404-b830-21ad196f78b5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.930412] env[62569]: DEBUG oslo_concurrency.lockutils [req-dc8563aa-87ef-4102-95c7-e9f4fee9d7f7 req-cd8d18af-b4c4-4e59-a8b4-84d5a7f4954a service nova] Lock "31ce29fa-4f60-4404-b830-21ad196f78b5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.930632] env[62569]: DEBUG oslo_concurrency.lockutils [req-dc8563aa-87ef-4102-95c7-e9f4fee9d7f7 req-cd8d18af-b4c4-4e59-a8b4-84d5a7f4954a service nova] Lock "31ce29fa-4f60-4404-b830-21ad196f78b5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.934026] env[62569]: DEBUG nova.compute.manager [req-dc8563aa-87ef-4102-95c7-e9f4fee9d7f7 req-cd8d18af-b4c4-4e59-a8b4-84d5a7f4954a service nova] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] No waiting events found dispatching network-vif-plugged-8e55dca6-f777-4333-9b49-feb5527d9a5e {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 852.934026] env[62569]: WARNING nova.compute.manager [req-dc8563aa-87ef-4102-95c7-e9f4fee9d7f7 req-cd8d18af-b4c4-4e59-a8b4-84d5a7f4954a service nova] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Received unexpected event network-vif-plugged-8e55dca6-f777-4333-9b49-feb5527d9a5e for instance with vm_state building and task_state spawning. [ 852.934026] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6caea880-59b3-49d1-a6fe-c1a4686b169d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.953130] env[62569]: DEBUG nova.compute.provider_tree [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 852.956282] env[62569]: DEBUG nova.compute.manager [req-11748069-9a1b-474b-9e44-fe4e0712ac54 req-375479be-9ad3-4e27-bc1b-91d4848676f8 service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Received event network-changed-96148207-1e82-4a26-baa5-65f88f889f97 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 852.956409] env[62569]: DEBUG nova.compute.manager [req-11748069-9a1b-474b-9e44-fe4e0712ac54 req-375479be-9ad3-4e27-bc1b-91d4848676f8 service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Refreshing instance network info cache due to event network-changed-96148207-1e82-4a26-baa5-65f88f889f97. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 852.957550] env[62569]: DEBUG oslo_concurrency.lockutils [req-11748069-9a1b-474b-9e44-fe4e0712ac54 req-375479be-9ad3-4e27-bc1b-91d4848676f8 service nova] Acquiring lock "refresh_cache-a5b3c345-149b-4dc3-84ff-fe72183efadb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.957550] env[62569]: DEBUG oslo_concurrency.lockutils [req-11748069-9a1b-474b-9e44-fe4e0712ac54 req-375479be-9ad3-4e27-bc1b-91d4848676f8 service nova] Acquired lock "refresh_cache-a5b3c345-149b-4dc3-84ff-fe72183efadb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.957550] env[62569]: DEBUG nova.network.neutron [req-11748069-9a1b-474b-9e44-fe4e0712ac54 req-375479be-9ad3-4e27-bc1b-91d4848676f8 service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Refreshing network info cache for port 96148207-1e82-4a26-baa5-65f88f889f97 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 853.172521] env[62569]: DEBUG oslo_vmware.api [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250203, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.238257] env[62569]: DEBUG nova.objects.base [None req-dccb5fb1-ce24-4e5e-af94-487a7aee6c09 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Object Instance<807d3025-d6a7-4778-a829-a61e2c7495c9> lazy-loaded attributes: flavor,pci_requests {{(pid=62569) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 853.238257] env[62569]: DEBUG nova.network.neutron [None req-dccb5fb1-ce24-4e5e-af94-487a7aee6c09 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 853.271106] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 853.271106] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b384ed2c-95e2-4618-b92b-50753a22bb11 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.278020] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 853.278020] env[62569]: value = "task-1250206" [ 853.278020] env[62569]: _type = "Task" [ 853.278020] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.285957] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] VM already powered off {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 853.286395] env[62569]: DEBUG nova.compute.manager [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 853.290208] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e637c511-53dc-488f-ae2e-da7022d41bcd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.294356] env[62569]: DEBUG oslo_concurrency.lockutils [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquiring lock "refresh_cache-337d88a2-30b9-4846-929e-042bd7a64a65" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.294356] env[62569]: DEBUG oslo_concurrency.lockutils [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquired lock "refresh_cache-337d88a2-30b9-4846-929e-042bd7a64a65" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.294356] env[62569]: DEBUG nova.network.neutron [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 853.345994] env[62569]: DEBUG nova.network.neutron [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 853.372396] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1a0d2531-4940-49b1-b1c7-b82f2d911bad tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Releasing lock "refresh_cache-a0b406c3-9466-41bd-9de1-e675cab2ceef" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.404901] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dccb5fb1-ce24-4e5e-af94-487a7aee6c09 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "interface-807d3025-d6a7-4778-a829-a61e2c7495c9-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.176s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.497855] env[62569]: DEBUG nova.scheduler.client.report [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Updated inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with generation 89 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 853.498223] env[62569]: DEBUG nova.compute.provider_tree [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Updating resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 generation from 89 to 90 during operation: update_inventory {{(pid=62569) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 853.498465] env[62569]: DEBUG nova.compute.provider_tree [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 853.562749] env[62569]: DEBUG nova.network.neutron [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Updating instance_info_cache with network_info: [{"id": "8e55dca6-f777-4333-9b49-feb5527d9a5e", "address": "fa:16:3e:97:39:72", "network": {"id": "c61b7867-7897-43ae-8f5b-19834c4390ea", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-512066189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ba7e96d226942bd99ab29fe703c6fea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e55dca6-f7", "ovs_interfaceid": "8e55dca6-f777-4333-9b49-feb5527d9a5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.672494] env[62569]: DEBUG oslo_vmware.api [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250203, 'name': CreateSnapshot_Task, 'duration_secs': 1.334754} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.672880] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Created Snapshot of the VM instance {{(pid=62569) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 853.673755] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0466a146-2581-46c5-a66e-883b6187aeda {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.694901] env[62569]: DEBUG nova.network.neutron [-] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.785202] env[62569]: DEBUG nova.network.neutron [req-11748069-9a1b-474b-9e44-fe4e0712ac54 req-375479be-9ad3-4e27-bc1b-91d4848676f8 service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Updated VIF entry in instance network info cache for port 96148207-1e82-4a26-baa5-65f88f889f97. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 853.785538] env[62569]: DEBUG nova.network.neutron [req-11748069-9a1b-474b-9e44-fe4e0712ac54 req-375479be-9ad3-4e27-bc1b-91d4848676f8 service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Updating instance_info_cache with network_info: [{"id": "96148207-1e82-4a26-baa5-65f88f889f97", "address": "fa:16:3e:03:33:1e", "network": {"id": "68456414-139e-4c9b-928c-1eba44e74d7f", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1573053904-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f881ca8a45234848aba4416ff1d582f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96148207-1e", "ovs_interfaceid": "96148207-1e82-4a26-baa5-65f88f889f97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.003994] env[62569]: DEBUG oslo_concurrency.lockutils [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.404s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.004791] env[62569]: DEBUG nova.compute.manager [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 854.012017] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.645s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.012017] env[62569]: DEBUG nova.objects.instance [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Lazy-loading 'resources' on Instance uuid 6e83fd32-215a-42e4-99e7-72bdce25555e {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 854.065263] env[62569]: DEBUG oslo_concurrency.lockutils [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Releasing lock "refresh_cache-31ce29fa-4f60-4404-b830-21ad196f78b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.065572] env[62569]: DEBUG nova.compute.manager [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Instance network_info: |[{"id": "8e55dca6-f777-4333-9b49-feb5527d9a5e", "address": "fa:16:3e:97:39:72", "network": {"id": "c61b7867-7897-43ae-8f5b-19834c4390ea", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-512066189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ba7e96d226942bd99ab29fe703c6fea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e55dca6-f7", "ovs_interfaceid": "8e55dca6-f777-4333-9b49-feb5527d9a5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 854.065987] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:97:39:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6685c85e-be1e-4b7b-a6cc-3e50e59b6567', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8e55dca6-f777-4333-9b49-feb5527d9a5e', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 854.076567] env[62569]: DEBUG oslo.service.loopingcall [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.077090] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 854.077329] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c7e6f06e-6aae-4a88-906c-46cfd0f0a1de {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.100210] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 854.100210] env[62569]: value = "task-1250207" [ 854.100210] env[62569]: _type = "Task" [ 854.100210] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.118574] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250207, 'name': CreateVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.175492] env[62569]: DEBUG nova.network.neutron [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Updating instance_info_cache with network_info: [{"id": "303f46d4-6e0a-418c-b62b-17323ab645b6", "address": "fa:16:3e:68:ca:b8", "network": {"id": "497ea2f3-6685-4d07-9952-b5ccf3c0b2f8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-561818925-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c69ac971dbd419f870a9fa68850cb1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7654928b-7afe-42e3-a18d-68ecc775cefe", "external-id": "cl2-zone-807", "segmentation_id": 807, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap303f46d4-6e", "ovs_interfaceid": "303f46d4-6e0a-418c-b62b-17323ab645b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.192566] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Creating linked-clone VM from snapshot {{(pid=62569) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 854.193741] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-57de8add-ddfd-48b4-b215-80988d17063d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.199063] env[62569]: INFO nova.compute.manager [-] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Took 1.28 seconds to deallocate network for instance. [ 854.205434] env[62569]: DEBUG oslo_vmware.api [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 854.205434] env[62569]: value = "task-1250208" [ 854.205434] env[62569]: _type = "Task" [ 854.205434] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.215220] env[62569]: DEBUG oslo_vmware.api [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250208, 'name': CloneVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.289039] env[62569]: DEBUG oslo_concurrency.lockutils [req-11748069-9a1b-474b-9e44-fe4e0712ac54 req-375479be-9ad3-4e27-bc1b-91d4848676f8 service nova] Releasing lock "refresh_cache-a5b3c345-149b-4dc3-84ff-fe72183efadb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.382025] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a0d2531-4940-49b1-b1c7-b82f2d911bad tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 854.382025] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-636d8a92-f79a-439c-9540-40b3dc3a94fc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.389270] env[62569]: DEBUG oslo_vmware.api [None req-1a0d2531-4940-49b1-b1c7-b82f2d911bad tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 854.389270] env[62569]: value = "task-1250209" [ 854.389270] env[62569]: _type = "Task" [ 854.389270] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.402295] env[62569]: DEBUG oslo_vmware.api [None req-1a0d2531-4940-49b1-b1c7-b82f2d911bad tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250209, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.512245] env[62569]: DEBUG nova.compute.utils [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 854.518858] env[62569]: DEBUG nova.compute.manager [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 854.519884] env[62569]: DEBUG nova.network.neutron [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 854.592680] env[62569]: DEBUG nova.policy [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e131aa363bb4b99801181b7e9ae8ec9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'edea4041206a4fd1bec5b7b62e0922bc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 854.614697] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250207, 'name': CreateVM_Task, 'duration_secs': 0.415646} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.615048] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 854.616163] env[62569]: DEBUG oslo_concurrency.lockutils [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.618017] env[62569]: DEBUG oslo_concurrency.lockutils [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.618017] env[62569]: DEBUG oslo_concurrency.lockutils [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 854.618017] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71e14c76-bd28-4035-adb7-b47d9ca029cf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.627456] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 854.627456] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ab2364-ce81-a6ac-e0b6-7cb1546e6d48" [ 854.627456] env[62569]: _type = "Task" [ 854.627456] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.638863] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ab2364-ce81-a6ac-e0b6-7cb1546e6d48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.681318] env[62569]: DEBUG oslo_concurrency.lockutils [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Releasing lock "refresh_cache-337d88a2-30b9-4846-929e-042bd7a64a65" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.705590] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.716625] env[62569]: DEBUG oslo_vmware.api [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250208, 'name': CloneVM_Task} progress is 94%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.784037] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1387b0c-212a-49e7-b5f8-8ede7556706f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.794729] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5d92137-6997-487f-bf14-ff0fc2a2c5d8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.835041] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f10fcec2-44d3-41f2-b9f3-985a780213e9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.843101] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffce99d2-7b4f-4bef-b9cf-e0e026507351 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.856656] env[62569]: DEBUG nova.compute.provider_tree [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.902922] env[62569]: DEBUG oslo_vmware.api [None req-1a0d2531-4940-49b1-b1c7-b82f2d911bad tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250209, 'name': PowerOnVM_Task, 'duration_secs': 0.411546} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.902922] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a0d2531-4940-49b1-b1c7-b82f2d911bad tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 854.902922] env[62569]: DEBUG nova.compute.manager [None req-1a0d2531-4940-49b1-b1c7-b82f2d911bad tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 854.902922] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f148133-38ed-492e-a16e-211d28690399 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.019684] env[62569]: DEBUG nova.compute.manager [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 855.100543] env[62569]: DEBUG nova.network.neutron [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Successfully created port: 0edf3387-b21d-47f6-b9e7-503d285ba554 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 855.136968] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ab2364-ce81-a6ac-e0b6-7cb1546e6d48, 'name': SearchDatastore_Task, 'duration_secs': 0.010732} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.137546] env[62569]: DEBUG oslo_concurrency.lockutils [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.137795] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 855.138052] env[62569]: DEBUG oslo_concurrency.lockutils [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.138207] env[62569]: DEBUG oslo_concurrency.lockutils [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.138395] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 855.138675] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6f4e4234-f1dc-4d39-93a8-9fbe235436d5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.146925] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 855.147045] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 855.147753] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81cf4642-5adc-4f6b-bc49-5f2d4089aeac {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.153263] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 855.153263] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5200ba7f-5b07-4421-7536-4a68d3719ec5" [ 855.153263] env[62569]: _type = "Task" [ 855.153263] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.161832] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5200ba7f-5b07-4421-7536-4a68d3719ec5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.216430] env[62569]: DEBUG oslo_vmware.api [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250208, 'name': CloneVM_Task} progress is 94%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.362326] env[62569]: DEBUG nova.scheduler.client.report [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 855.665986] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5200ba7f-5b07-4421-7536-4a68d3719ec5, 'name': SearchDatastore_Task, 'duration_secs': 0.011402} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.666832] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8a74984-2b87-40a4-8be5-3331ccc3e062 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.676458] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 855.676458] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52fef334-1308-d8a6-4943-7e8940dbb8c3" [ 855.676458] env[62569]: _type = "Task" [ 855.676458] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.687290] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52fef334-1308-d8a6-4943-7e8940dbb8c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.723377] env[62569]: DEBUG oslo_vmware.api [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250208, 'name': CloneVM_Task} progress is 95%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.851137] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 855.852744] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ff3e1e2-2f45-4b31-be2a-cd86e61d92a8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.860562] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 855.861906] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cf3e7fd9-7af0-4050-877e-9affe1c2f9d3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.865215] env[62569]: DEBUG nova.compute.manager [req-65f5c5c7-403f-416e-9cc0-e94c8d59f9b5 req-cb518cd0-0bf4-4c38-928e-c5c885ddd950 service nova] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Received event network-changed-8e55dca6-f777-4333-9b49-feb5527d9a5e {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 855.865367] env[62569]: DEBUG nova.compute.manager [req-65f5c5c7-403f-416e-9cc0-e94c8d59f9b5 req-cb518cd0-0bf4-4c38-928e-c5c885ddd950 service nova] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Refreshing instance network info cache due to event network-changed-8e55dca6-f777-4333-9b49-feb5527d9a5e. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 855.865581] env[62569]: DEBUG oslo_concurrency.lockutils [req-65f5c5c7-403f-416e-9cc0-e94c8d59f9b5 req-cb518cd0-0bf4-4c38-928e-c5c885ddd950 service nova] Acquiring lock "refresh_cache-31ce29fa-4f60-4404-b830-21ad196f78b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.865734] env[62569]: DEBUG oslo_concurrency.lockutils [req-65f5c5c7-403f-416e-9cc0-e94c8d59f9b5 req-cb518cd0-0bf4-4c38-928e-c5c885ddd950 service nova] Acquired lock "refresh_cache-31ce29fa-4f60-4404-b830-21ad196f78b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.865900] env[62569]: DEBUG nova.network.neutron [req-65f5c5c7-403f-416e-9cc0-e94c8d59f9b5 req-cb518cd0-0bf4-4c38-928e-c5c885ddd950 service nova] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Refreshing network info cache for port 8e55dca6-f777-4333-9b49-feb5527d9a5e {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 855.870021] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.860s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.870896] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.763s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.871167] env[62569]: DEBUG nova.objects.instance [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lazy-loading 'resources' on Instance uuid c9568fcb-745f-4c04-8487-a38a617aab7c {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 855.882483] env[62569]: DEBUG nova.compute.manager [req-c58dcbc4-a656-40af-b03a-b30aace983d2 req-2fd6cf7a-586d-40a6-b6e8-af96570bc1f9 service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Received event network-changed-96148207-1e82-4a26-baa5-65f88f889f97 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 855.882483] env[62569]: DEBUG nova.compute.manager [req-c58dcbc4-a656-40af-b03a-b30aace983d2 req-2fd6cf7a-586d-40a6-b6e8-af96570bc1f9 service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Refreshing instance network info cache due to event network-changed-96148207-1e82-4a26-baa5-65f88f889f97. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 855.882483] env[62569]: DEBUG oslo_concurrency.lockutils [req-c58dcbc4-a656-40af-b03a-b30aace983d2 req-2fd6cf7a-586d-40a6-b6e8-af96570bc1f9 service nova] Acquiring lock "refresh_cache-a5b3c345-149b-4dc3-84ff-fe72183efadb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.882483] env[62569]: DEBUG oslo_concurrency.lockutils [req-c58dcbc4-a656-40af-b03a-b30aace983d2 req-2fd6cf7a-586d-40a6-b6e8-af96570bc1f9 service nova] Acquired lock "refresh_cache-a5b3c345-149b-4dc3-84ff-fe72183efadb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.882483] env[62569]: DEBUG nova.network.neutron [req-c58dcbc4-a656-40af-b03a-b30aace983d2 req-2fd6cf7a-586d-40a6-b6e8-af96570bc1f9 service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Refreshing network info cache for port 96148207-1e82-4a26-baa5-65f88f889f97 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 855.902686] env[62569]: INFO nova.scheduler.client.report [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Deleted allocations for instance 6e83fd32-215a-42e4-99e7-72bdce25555e [ 855.941108] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 855.941108] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 855.941108] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Deleting the datastore file [datastore2] 337d88a2-30b9-4846-929e-042bd7a64a65 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 855.941108] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3406fe44-8853-4c30-8983-f5811f743762 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.949270] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 855.949270] env[62569]: value = "task-1250211" [ 855.949270] env[62569]: _type = "Task" [ 855.949270] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.961199] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250211, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.034138] env[62569]: DEBUG nova.compute.manager [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 856.063456] env[62569]: DEBUG nova.virt.hardware [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 856.063774] env[62569]: DEBUG nova.virt.hardware [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 856.063989] env[62569]: DEBUG nova.virt.hardware [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 856.064159] env[62569]: DEBUG nova.virt.hardware [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 856.064416] env[62569]: DEBUG nova.virt.hardware [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 856.064672] env[62569]: DEBUG nova.virt.hardware [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 856.064945] env[62569]: DEBUG nova.virt.hardware [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 856.065162] env[62569]: DEBUG nova.virt.hardware [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 856.065358] env[62569]: DEBUG nova.virt.hardware [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 856.065560] env[62569]: DEBUG nova.virt.hardware [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 856.065767] env[62569]: DEBUG nova.virt.hardware [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 856.066660] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e170f10c-d81e-41b3-964b-42817f627f41 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.075185] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-285014a0-8b83-46a6-803c-b6e38dc6642f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.128229] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "interface-807d3025-d6a7-4778-a829-a61e2c7495c9-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.128455] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "interface-807d3025-d6a7-4778-a829-a61e2c7495c9-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.129024] env[62569]: DEBUG nova.objects.instance [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lazy-loading 'flavor' on Instance uuid 807d3025-d6a7-4778-a829-a61e2c7495c9 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 856.937570] env[62569]: DEBUG nova.network.neutron [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Successfully updated port: 0edf3387-b21d-47f6-b9e7-503d285ba554 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 856.951794] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b20e51c2-6d3e-4f59-b760-458bf8fa8371 tempest-ServerGroupTestJSON-1324971923 tempest-ServerGroupTestJSON-1324971923-project-member] Lock "6e83fd32-215a-42e4-99e7-72bdce25555e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.080s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.968891] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52fef334-1308-d8a6-4943-7e8940dbb8c3, 'name': SearchDatastore_Task, 'duration_secs': 0.020561} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.976928] env[62569]: DEBUG oslo_concurrency.lockutils [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.976928] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 31ce29fa-4f60-4404-b830-21ad196f78b5/31ce29fa-4f60-4404-b830-21ad196f78b5.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 856.976928] env[62569]: DEBUG oslo_vmware.api [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250208, 'name': CloneVM_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.977314] env[62569]: DEBUG oslo_vmware.api [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250211, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.398192} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.977545] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-11337bd5-402a-43e2-9448-f50c18ebedc8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.980622] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 856.980622] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 856.980622] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 856.994488] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 856.994488] env[62569]: value = "task-1250212" [ 856.994488] env[62569]: _type = "Task" [ 856.994488] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.007200] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250212, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.013371] env[62569]: INFO nova.scheduler.client.report [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Deleted allocations for instance 337d88a2-30b9-4846-929e-042bd7a64a65 [ 857.154360] env[62569]: DEBUG nova.objects.instance [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lazy-loading 'pci_requests' on Instance uuid 807d3025-d6a7-4778-a829-a61e2c7495c9 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 857.156594] env[62569]: DEBUG oslo_concurrency.lockutils [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Acquiring lock "298efba9-226d-4105-8b32-76b3dddddb0b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.156846] env[62569]: DEBUG oslo_concurrency.lockutils [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Lock "298efba9-226d-4105-8b32-76b3dddddb0b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.219273] env[62569]: DEBUG nova.network.neutron [req-c58dcbc4-a656-40af-b03a-b30aace983d2 req-2fd6cf7a-586d-40a6-b6e8-af96570bc1f9 service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Updated VIF entry in instance network info cache for port 96148207-1e82-4a26-baa5-65f88f889f97. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 857.219881] env[62569]: DEBUG nova.network.neutron [req-c58dcbc4-a656-40af-b03a-b30aace983d2 req-2fd6cf7a-586d-40a6-b6e8-af96570bc1f9 service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Updating instance_info_cache with network_info: [{"id": "96148207-1e82-4a26-baa5-65f88f889f97", "address": "fa:16:3e:03:33:1e", "network": {"id": "68456414-139e-4c9b-928c-1eba44e74d7f", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1573053904-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f881ca8a45234848aba4416ff1d582f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96148207-1e", "ovs_interfaceid": "96148207-1e82-4a26-baa5-65f88f889f97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.238811] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72cb04d9-f756-4bb1-b0b2-2bf2a33d24c0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.248549] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b67c844-251c-4ebc-af59-f94ea927e2d6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.283185] env[62569]: DEBUG nova.network.neutron [req-65f5c5c7-403f-416e-9cc0-e94c8d59f9b5 req-cb518cd0-0bf4-4c38-928e-c5c885ddd950 service nova] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Updated VIF entry in instance network info cache for port 8e55dca6-f777-4333-9b49-feb5527d9a5e. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 857.283633] env[62569]: DEBUG nova.network.neutron [req-65f5c5c7-403f-416e-9cc0-e94c8d59f9b5 req-cb518cd0-0bf4-4c38-928e-c5c885ddd950 service nova] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Updating instance_info_cache with network_info: [{"id": "8e55dca6-f777-4333-9b49-feb5527d9a5e", "address": "fa:16:3e:97:39:72", "network": {"id": "c61b7867-7897-43ae-8f5b-19834c4390ea", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-512066189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ba7e96d226942bd99ab29fe703c6fea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e55dca6-f7", "ovs_interfaceid": "8e55dca6-f777-4333-9b49-feb5527d9a5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.286512] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94f9a611-115b-4b16-8853-8de60ffb65d7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.299789] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b6fda43-37b9-4ca4-9195-fb34f7b5e088 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.316622] env[62569]: DEBUG nova.compute.provider_tree [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.450722] env[62569]: DEBUG oslo_vmware.api [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250208, 'name': CloneVM_Task, 'duration_secs': 2.847942} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.451227] env[62569]: INFO nova.virt.vmwareapi.vmops [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Created linked-clone VM from snapshot [ 857.451966] env[62569]: DEBUG oslo_concurrency.lockutils [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Acquiring lock "refresh_cache-691d8fbe-b9ee-454f-bd7b-14520e53ed26" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.452803] env[62569]: DEBUG oslo_concurrency.lockutils [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Acquired lock "refresh_cache-691d8fbe-b9ee-454f-bd7b-14520e53ed26" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.452803] env[62569]: DEBUG nova.network.neutron [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 857.454652] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e83916d-eecd-44ff-be1b-65065c0ff5da {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.465293] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Uploading image 5909e6a6-6c00-4956-a4d0-c6430d91b165 {{(pid=62569) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 857.493235] env[62569]: DEBUG oslo_vmware.rw_handles [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 857.493235] env[62569]: value = "vm-269439" [ 857.493235] env[62569]: _type = "VirtualMachine" [ 857.493235] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 857.493626] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e16da847-e7a5-46e9-8790-a9ae1a3a5f8e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.507288] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250212, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.509212] env[62569]: DEBUG oslo_vmware.rw_handles [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lease: (returnval){ [ 857.509212] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520efa9b-3fe8-3192-858b-4a3b00c4accb" [ 857.509212] env[62569]: _type = "HttpNfcLease" [ 857.509212] env[62569]: } obtained for exporting VM: (result){ [ 857.509212] env[62569]: value = "vm-269439" [ 857.509212] env[62569]: _type = "VirtualMachine" [ 857.509212] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 857.509572] env[62569]: DEBUG oslo_vmware.api [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the lease: (returnval){ [ 857.509572] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520efa9b-3fe8-3192-858b-4a3b00c4accb" [ 857.509572] env[62569]: _type = "HttpNfcLease" [ 857.509572] env[62569]: } to be ready. {{(pid=62569) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 857.517434] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 857.517434] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520efa9b-3fe8-3192-858b-4a3b00c4accb" [ 857.517434] env[62569]: _type = "HttpNfcLease" [ 857.517434] env[62569]: } is initializing. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 857.522068] env[62569]: DEBUG oslo_concurrency.lockutils [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.659441] env[62569]: DEBUG nova.objects.base [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Object Instance<807d3025-d6a7-4778-a829-a61e2c7495c9> lazy-loaded attributes: flavor,pci_requests {{(pid=62569) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 857.659839] env[62569]: DEBUG nova.network.neutron [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 857.662504] env[62569]: DEBUG nova.compute.manager [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 857.704038] env[62569]: DEBUG nova.policy [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ab6d20862c54432cbafdda33ccc974ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '530ca7e7924743ab91a362a064a3111b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 857.727984] env[62569]: DEBUG oslo_concurrency.lockutils [req-c58dcbc4-a656-40af-b03a-b30aace983d2 req-2fd6cf7a-586d-40a6-b6e8-af96570bc1f9 service nova] Releasing lock "refresh_cache-a5b3c345-149b-4dc3-84ff-fe72183efadb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.728350] env[62569]: DEBUG nova.compute.manager [req-c58dcbc4-a656-40af-b03a-b30aace983d2 req-2fd6cf7a-586d-40a6-b6e8-af96570bc1f9 service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Received event network-changed-96148207-1e82-4a26-baa5-65f88f889f97 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 857.728533] env[62569]: DEBUG nova.compute.manager [req-c58dcbc4-a656-40af-b03a-b30aace983d2 req-2fd6cf7a-586d-40a6-b6e8-af96570bc1f9 service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Refreshing instance network info cache due to event network-changed-96148207-1e82-4a26-baa5-65f88f889f97. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 857.728754] env[62569]: DEBUG oslo_concurrency.lockutils [req-c58dcbc4-a656-40af-b03a-b30aace983d2 req-2fd6cf7a-586d-40a6-b6e8-af96570bc1f9 service nova] Acquiring lock "refresh_cache-a5b3c345-149b-4dc3-84ff-fe72183efadb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.729160] env[62569]: DEBUG oslo_concurrency.lockutils [req-c58dcbc4-a656-40af-b03a-b30aace983d2 req-2fd6cf7a-586d-40a6-b6e8-af96570bc1f9 service nova] Acquired lock "refresh_cache-a5b3c345-149b-4dc3-84ff-fe72183efadb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.729234] env[62569]: DEBUG nova.network.neutron [req-c58dcbc4-a656-40af-b03a-b30aace983d2 req-2fd6cf7a-586d-40a6-b6e8-af96570bc1f9 service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Refreshing network info cache for port 96148207-1e82-4a26-baa5-65f88f889f97 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 857.794956] env[62569]: DEBUG oslo_concurrency.lockutils [req-65f5c5c7-403f-416e-9cc0-e94c8d59f9b5 req-cb518cd0-0bf4-4c38-928e-c5c885ddd950 service nova] Releasing lock "refresh_cache-31ce29fa-4f60-4404-b830-21ad196f78b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.795117] env[62569]: DEBUG nova.compute.manager [req-65f5c5c7-403f-416e-9cc0-e94c8d59f9b5 req-cb518cd0-0bf4-4c38-928e-c5c885ddd950 service nova] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Received event network-vif-deleted-030d3c5e-456b-4a2d-a229-29d169db3996 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 857.795770] env[62569]: INFO nova.compute.manager [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Rebuilding instance [ 857.820305] env[62569]: DEBUG nova.scheduler.client.report [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 857.856693] env[62569]: DEBUG nova.compute.manager [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 857.857523] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ec15e27-c29e-4226-b390-3a985155f8d8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.973613] env[62569]: DEBUG nova.network.neutron [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Successfully created port: 5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 858.006763] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250212, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.646706} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.007433] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 31ce29fa-4f60-4404-b830-21ad196f78b5/31ce29fa-4f60-4404-b830-21ad196f78b5.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 858.007706] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 858.007976] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d394eec1-2887-44f8-b4bf-68e7e404a19d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.019139] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 858.019139] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520efa9b-3fe8-3192-858b-4a3b00c4accb" [ 858.019139] env[62569]: _type = "HttpNfcLease" [ 858.019139] env[62569]: } is ready. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 858.020408] env[62569]: DEBUG oslo_vmware.rw_handles [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 858.020408] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520efa9b-3fe8-3192-858b-4a3b00c4accb" [ 858.020408] env[62569]: _type = "HttpNfcLease" [ 858.020408] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 858.020714] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 858.020714] env[62569]: value = "task-1250214" [ 858.020714] env[62569]: _type = "Task" [ 858.020714] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.021744] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4db8b89-df11-445c-9ab4-12dc63f88226 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.026015] env[62569]: DEBUG nova.network.neutron [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 858.037295] env[62569]: DEBUG oslo_vmware.rw_handles [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5237f662-615a-a9d3-b3df-ef2a6b53a3ea/disk-0.vmdk from lease info. {{(pid=62569) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 858.037508] env[62569]: DEBUG oslo_vmware.rw_handles [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5237f662-615a-a9d3-b3df-ef2a6b53a3ea/disk-0.vmdk for reading. {{(pid=62569) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 858.042318] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250214, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.138432] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Acquiring lock "a5b3c345-149b-4dc3-84ff-fe72183efadb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.138902] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Lock "a5b3c345-149b-4dc3-84ff-fe72183efadb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.139243] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Acquiring lock "a5b3c345-149b-4dc3-84ff-fe72183efadb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.139469] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Lock "a5b3c345-149b-4dc3-84ff-fe72183efadb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.139654] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Lock "a5b3c345-149b-4dc3-84ff-fe72183efadb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.142021] env[62569]: INFO nova.compute.manager [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Terminating instance [ 858.187518] env[62569]: DEBUG oslo_concurrency.lockutils [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.280608] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c962310b-2b20-4af4-bd76-bf4b12888b33 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.327913] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.457s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.329906] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 6.622s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.353984] env[62569]: INFO nova.scheduler.client.report [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Deleted allocations for instance c9568fcb-745f-4c04-8487-a38a617aab7c [ 858.371357] env[62569]: DEBUG nova.network.neutron [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Updating instance_info_cache with network_info: [{"id": "0edf3387-b21d-47f6-b9e7-503d285ba554", "address": "fa:16:3e:cb:d0:d3", "network": {"id": "e67c1ed0-2a04-4a64-9041-c9dbe857c504", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-566826463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "edea4041206a4fd1bec5b7b62e0922bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0edf3387-b2", "ovs_interfaceid": "0edf3387-b21d-47f6-b9e7-503d285ba554", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.534656] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250214, 'name': ExtendVirtualDisk_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.561529] env[62569]: DEBUG nova.compute.manager [req-9e28a29a-c428-4dd6-96c2-0fa1956c07de req-6fc4f7fc-9b16-4b86-8947-013bf0de5866 service nova] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Received event network-vif-plugged-0edf3387-b21d-47f6-b9e7-503d285ba554 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 858.561746] env[62569]: DEBUG oslo_concurrency.lockutils [req-9e28a29a-c428-4dd6-96c2-0fa1956c07de req-6fc4f7fc-9b16-4b86-8947-013bf0de5866 service nova] Acquiring lock "691d8fbe-b9ee-454f-bd7b-14520e53ed26-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.561981] env[62569]: DEBUG oslo_concurrency.lockutils [req-9e28a29a-c428-4dd6-96c2-0fa1956c07de req-6fc4f7fc-9b16-4b86-8947-013bf0de5866 service nova] Lock "691d8fbe-b9ee-454f-bd7b-14520e53ed26-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.563126] env[62569]: DEBUG oslo_concurrency.lockutils [req-9e28a29a-c428-4dd6-96c2-0fa1956c07de req-6fc4f7fc-9b16-4b86-8947-013bf0de5866 service nova] Lock "691d8fbe-b9ee-454f-bd7b-14520e53ed26-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.563126] env[62569]: DEBUG nova.compute.manager [req-9e28a29a-c428-4dd6-96c2-0fa1956c07de req-6fc4f7fc-9b16-4b86-8947-013bf0de5866 service nova] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] No waiting events found dispatching network-vif-plugged-0edf3387-b21d-47f6-b9e7-503d285ba554 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 858.563126] env[62569]: WARNING nova.compute.manager [req-9e28a29a-c428-4dd6-96c2-0fa1956c07de req-6fc4f7fc-9b16-4b86-8947-013bf0de5866 service nova] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Received unexpected event network-vif-plugged-0edf3387-b21d-47f6-b9e7-503d285ba554 for instance with vm_state building and task_state spawning. [ 858.563126] env[62569]: DEBUG nova.compute.manager [req-9e28a29a-c428-4dd6-96c2-0fa1956c07de req-6fc4f7fc-9b16-4b86-8947-013bf0de5866 service nova] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Received event network-changed-0edf3387-b21d-47f6-b9e7-503d285ba554 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 858.563126] env[62569]: DEBUG nova.compute.manager [req-9e28a29a-c428-4dd6-96c2-0fa1956c07de req-6fc4f7fc-9b16-4b86-8947-013bf0de5866 service nova] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Refreshing instance network info cache due to event network-changed-0edf3387-b21d-47f6-b9e7-503d285ba554. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 858.563379] env[62569]: DEBUG oslo_concurrency.lockutils [req-9e28a29a-c428-4dd6-96c2-0fa1956c07de req-6fc4f7fc-9b16-4b86-8947-013bf0de5866 service nova] Acquiring lock "refresh_cache-691d8fbe-b9ee-454f-bd7b-14520e53ed26" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.564272] env[62569]: DEBUG nova.network.neutron [req-c58dcbc4-a656-40af-b03a-b30aace983d2 req-2fd6cf7a-586d-40a6-b6e8-af96570bc1f9 service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Updated VIF entry in instance network info cache for port 96148207-1e82-4a26-baa5-65f88f889f97. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 858.565053] env[62569]: DEBUG nova.network.neutron [req-c58dcbc4-a656-40af-b03a-b30aace983d2 req-2fd6cf7a-586d-40a6-b6e8-af96570bc1f9 service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Updating instance_info_cache with network_info: [{"id": "96148207-1e82-4a26-baa5-65f88f889f97", "address": "fa:16:3e:03:33:1e", "network": {"id": "68456414-139e-4c9b-928c-1eba44e74d7f", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1573053904-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "f881ca8a45234848aba4416ff1d582f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96148207-1e", "ovs_interfaceid": "96148207-1e82-4a26-baa5-65f88f889f97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.582146] env[62569]: DEBUG nova.compute.manager [req-9451814b-b513-4af1-94d2-88a8df9a63fc req-f03d80e5-be5c-4fd0-9b22-cf191cd3f33a service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Received event network-vif-unplugged-303f46d4-6e0a-418c-b62b-17323ab645b6 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 858.582401] env[62569]: DEBUG oslo_concurrency.lockutils [req-9451814b-b513-4af1-94d2-88a8df9a63fc req-f03d80e5-be5c-4fd0-9b22-cf191cd3f33a service nova] Acquiring lock "337d88a2-30b9-4846-929e-042bd7a64a65-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.582614] env[62569]: DEBUG oslo_concurrency.lockutils [req-9451814b-b513-4af1-94d2-88a8df9a63fc req-f03d80e5-be5c-4fd0-9b22-cf191cd3f33a service nova] Lock "337d88a2-30b9-4846-929e-042bd7a64a65-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.582791] env[62569]: DEBUG oslo_concurrency.lockutils [req-9451814b-b513-4af1-94d2-88a8df9a63fc req-f03d80e5-be5c-4fd0-9b22-cf191cd3f33a service nova] Lock "337d88a2-30b9-4846-929e-042bd7a64a65-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.582991] env[62569]: DEBUG nova.compute.manager [req-9451814b-b513-4af1-94d2-88a8df9a63fc req-f03d80e5-be5c-4fd0-9b22-cf191cd3f33a service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] No waiting events found dispatching network-vif-unplugged-303f46d4-6e0a-418c-b62b-17323ab645b6 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 858.583191] env[62569]: WARNING nova.compute.manager [req-9451814b-b513-4af1-94d2-88a8df9a63fc req-f03d80e5-be5c-4fd0-9b22-cf191cd3f33a service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Received unexpected event network-vif-unplugged-303f46d4-6e0a-418c-b62b-17323ab645b6 for instance with vm_state shelved_offloaded and task_state None. [ 858.583351] env[62569]: DEBUG nova.compute.manager [req-9451814b-b513-4af1-94d2-88a8df9a63fc req-f03d80e5-be5c-4fd0-9b22-cf191cd3f33a service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Received event network-changed-303f46d4-6e0a-418c-b62b-17323ab645b6 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 858.583527] env[62569]: DEBUG nova.compute.manager [req-9451814b-b513-4af1-94d2-88a8df9a63fc req-f03d80e5-be5c-4fd0-9b22-cf191cd3f33a service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Refreshing instance network info cache due to event network-changed-303f46d4-6e0a-418c-b62b-17323ab645b6. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 858.583715] env[62569]: DEBUG oslo_concurrency.lockutils [req-9451814b-b513-4af1-94d2-88a8df9a63fc req-f03d80e5-be5c-4fd0-9b22-cf191cd3f33a service nova] Acquiring lock "refresh_cache-337d88a2-30b9-4846-929e-042bd7a64a65" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.583854] env[62569]: DEBUG oslo_concurrency.lockutils [req-9451814b-b513-4af1-94d2-88a8df9a63fc req-f03d80e5-be5c-4fd0-9b22-cf191cd3f33a service nova] Acquired lock "refresh_cache-337d88a2-30b9-4846-929e-042bd7a64a65" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.584042] env[62569]: DEBUG nova.network.neutron [req-9451814b-b513-4af1-94d2-88a8df9a63fc req-f03d80e5-be5c-4fd0-9b22-cf191cd3f33a service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Refreshing network info cache for port 303f46d4-6e0a-418c-b62b-17323ab645b6 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 858.647610] env[62569]: DEBUG nova.compute.manager [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 858.648625] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 858.649795] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09b5f7a2-24d4-4f7b-b513-fdb788d9e23c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.662054] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 858.662381] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a0ac1599-419c-4446-916c-469d1744a532 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.670545] env[62569]: DEBUG oslo_vmware.api [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Waiting for the task: (returnval){ [ 858.670545] env[62569]: value = "task-1250215" [ 858.670545] env[62569]: _type = "Task" [ 858.670545] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.680687] env[62569]: DEBUG oslo_vmware.api [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250215, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.870893] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 858.870893] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9661a2dd-9ab4-4e6d-aeab-d053252fb691 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "c9568fcb-745f-4c04-8487-a38a617aab7c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.863s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.871515] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fc39cd87-1067-4245-ac95-50a41c91b557 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.875553] env[62569]: DEBUG oslo_concurrency.lockutils [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Releasing lock "refresh_cache-691d8fbe-b9ee-454f-bd7b-14520e53ed26" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.875553] env[62569]: DEBUG nova.compute.manager [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Instance network_info: |[{"id": "0edf3387-b21d-47f6-b9e7-503d285ba554", "address": "fa:16:3e:cb:d0:d3", "network": {"id": "e67c1ed0-2a04-4a64-9041-c9dbe857c504", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-566826463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "edea4041206a4fd1bec5b7b62e0922bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0edf3387-b2", "ovs_interfaceid": "0edf3387-b21d-47f6-b9e7-503d285ba554", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 858.875553] env[62569]: DEBUG oslo_concurrency.lockutils [req-9e28a29a-c428-4dd6-96c2-0fa1956c07de req-6fc4f7fc-9b16-4b86-8947-013bf0de5866 service nova] Acquired lock "refresh_cache-691d8fbe-b9ee-454f-bd7b-14520e53ed26" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.875553] env[62569]: DEBUG nova.network.neutron [req-9e28a29a-c428-4dd6-96c2-0fa1956c07de req-6fc4f7fc-9b16-4b86-8947-013bf0de5866 service nova] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Refreshing network info cache for port 0edf3387-b21d-47f6-b9e7-503d285ba554 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 858.876197] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cb:d0:d3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24ec44b7-0acf-4ff9-8bb3-4641b74af7a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0edf3387-b21d-47f6-b9e7-503d285ba554', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 858.883972] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Creating folder: Project (edea4041206a4fd1bec5b7b62e0922bc). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 858.885423] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-919fb50e-cfb3-4996-8c26-4f63ecbf4f2f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.889878] env[62569]: DEBUG oslo_vmware.api [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Waiting for the task: (returnval){ [ 858.889878] env[62569]: value = "task-1250216" [ 858.889878] env[62569]: _type = "Task" [ 858.889878] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.896212] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Created folder: Project (edea4041206a4fd1bec5b7b62e0922bc) in parent group-v269330. [ 858.898685] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Creating folder: Instances. Parent ref: group-v269440. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 858.902500] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6c88907b-6dde-4cbe-a211-40446b531759 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.904535] env[62569]: DEBUG oslo_vmware.api [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250216, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.913492] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Created folder: Instances in parent group-v269440. [ 858.914693] env[62569]: DEBUG oslo.service.loopingcall [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 858.914693] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 858.914693] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-32bd9381-9080-421a-b854-f27e46d32ebc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.944024] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 858.944024] env[62569]: value = "task-1250219" [ 858.944024] env[62569]: _type = "Task" [ 858.944024] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.951140] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250219, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.035210] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250214, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.517107} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.035659] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 859.036652] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7520b7fa-7505-43c3-8e9c-0d6990705040 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.060604] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 31ce29fa-4f60-4404-b830-21ad196f78b5/31ce29fa-4f60-4404-b830-21ad196f78b5.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 859.061461] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd7e46d7-0999-4646-977c-8b9d26c885f6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.077488] env[62569]: DEBUG oslo_concurrency.lockutils [req-c58dcbc4-a656-40af-b03a-b30aace983d2 req-2fd6cf7a-586d-40a6-b6e8-af96570bc1f9 service nova] Releasing lock "refresh_cache-a5b3c345-149b-4dc3-84ff-fe72183efadb" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.090238] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 859.090238] env[62569]: value = "task-1250220" [ 859.090238] env[62569]: _type = "Task" [ 859.090238] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.100715] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250220, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.182227] env[62569]: DEBUG oslo_vmware.api [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250215, 'name': PowerOffVM_Task, 'duration_secs': 0.316581} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.182541] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 859.182978] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 859.182978] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-59d864a3-ee1f-4af0-823e-3a462e1f4375 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.282896] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 859.283260] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 859.283338] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Deleting the datastore file [datastore2] a5b3c345-149b-4dc3-84ff-fe72183efadb {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 859.283760] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e250dd00-0c92-4591-af34-4f56fb3013c0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.296659] env[62569]: DEBUG oslo_vmware.api [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Waiting for the task: (returnval){ [ 859.296659] env[62569]: value = "task-1250222" [ 859.296659] env[62569]: _type = "Task" [ 859.296659] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.309335] env[62569]: DEBUG oslo_vmware.api [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250222, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.377882] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance a0b406c3-9466-41bd-9de1-e675cab2ceef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 859.377882] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance b6e5eefc-8c06-445b-a3af-9404578b6179 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 859.377882] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 807d3025-d6a7-4778-a829-a61e2c7495c9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 859.377882] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance fdeac8b8-a34d-4b4c-8f72-281f84e22ebe actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 859.377882] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance b47cd2d7-0cd2-41af-8ed1-a6dfca323516 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 859.377882] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance cb9d8991-b7ba-4bf7-9d22-1391f4cea768 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 859.377882] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance e4eadcdf-a04a-4255-ba1c-fe20156c655f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 859.377882] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance a5b3c345-149b-4dc3-84ff-fe72183efadb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 859.377882] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 05db90a6-3faf-4878-b782-ed17df47ed5f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 859.378594] env[62569]: WARNING nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance c8edf395-ea21-4ba8-991e-0f43fe4ee830 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 859.378594] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 31ce29fa-4f60-4404-b830-21ad196f78b5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 859.378594] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 691d8fbe-b9ee-454f-bd7b-14520e53ed26 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 859.401864] env[62569]: DEBUG oslo_vmware.api [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250216, 'name': PowerOffVM_Task, 'duration_secs': 0.238437} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.402374] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 859.402937] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 859.403372] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6fecfe95-71f1-4b33-adbc-43e18d888d77 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.410111] env[62569]: DEBUG oslo_vmware.api [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Waiting for the task: (returnval){ [ 859.410111] env[62569]: value = "task-1250223" [ 859.410111] env[62569]: _type = "Task" [ 859.410111] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.423171] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] VM already powered off {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 859.423434] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Volume detach. Driver type: vmdk {{(pid=62569) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 859.423711] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269356', 'volume_id': '3fb6b4c3-cf28-44c4-9606-7430ec94b7f6', 'name': 'volume-3fb6b4c3-cf28-44c4-9606-7430ec94b7f6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'fdeac8b8-a34d-4b4c-8f72-281f84e22ebe', 'attached_at': '', 'detached_at': '', 'volume_id': '3fb6b4c3-cf28-44c4-9606-7430ec94b7f6', 'serial': '3fb6b4c3-cf28-44c4-9606-7430ec94b7f6'} {{(pid=62569) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 859.428416] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d55f36ca-44cb-41aa-8327-e27ff37c0621 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.431930] env[62569]: DEBUG nova.network.neutron [req-9451814b-b513-4af1-94d2-88a8df9a63fc req-f03d80e5-be5c-4fd0-9b22-cf191cd3f33a service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Updated VIF entry in instance network info cache for port 303f46d4-6e0a-418c-b62b-17323ab645b6. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 859.432465] env[62569]: DEBUG nova.network.neutron [req-9451814b-b513-4af1-94d2-88a8df9a63fc req-f03d80e5-be5c-4fd0-9b22-cf191cd3f33a service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Updating instance_info_cache with network_info: [{"id": "303f46d4-6e0a-418c-b62b-17323ab645b6", "address": "fa:16:3e:68:ca:b8", "network": {"id": "497ea2f3-6685-4d07-9952-b5ccf3c0b2f8", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-561818925-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c69ac971dbd419f870a9fa68850cb1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap303f46d4-6e", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.463270] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7abda70d-49ea-406b-8d54-6a5fe0b7e1b3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.479813] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250219, 'name': CreateVM_Task, 'duration_secs': 0.367361} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.482058] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 859.483179] env[62569]: DEBUG oslo_concurrency.lockutils [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.483476] env[62569]: DEBUG oslo_concurrency.lockutils [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.483929] env[62569]: DEBUG oslo_concurrency.lockutils [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 859.485860] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a719cf8-c2a2-41cc-894d-9578d8728937 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.491665] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5cf9eff-1c7b-4c4a-9af4-c80090aa55fd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.513173] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c854cdb-01f7-41b6-a402-49549107e3cc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.518291] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Waiting for the task: (returnval){ [ 859.518291] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]523c34c8-de32-c5fe-2143-2fb730d7e5a0" [ 859.518291] env[62569]: _type = "Task" [ 859.518291] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.533673] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] The volume has not been displaced from its original location: [datastore1] volume-3fb6b4c3-cf28-44c4-9606-7430ec94b7f6/volume-3fb6b4c3-cf28-44c4-9606-7430ec94b7f6.vmdk. No consolidation needed. {{(pid=62569) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 859.538973] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Reconfiguring VM instance instance-00000046 to detach disk 2000 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 859.539883] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9596dca-94d9-4f97-a5d4-762296e5ed67 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.557818] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]523c34c8-de32-c5fe-2143-2fb730d7e5a0, 'name': SearchDatastore_Task, 'duration_secs': 0.01989} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.558744] env[62569]: DEBUG oslo_concurrency.lockutils [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.559000] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 859.559367] env[62569]: DEBUG oslo_concurrency.lockutils [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.559577] env[62569]: DEBUG oslo_concurrency.lockutils [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.559833] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 859.560191] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a5ee0cb3-e1f8-48e1-ac98-1ac3cff17a3c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.563446] env[62569]: DEBUG oslo_vmware.api [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Waiting for the task: (returnval){ [ 859.563446] env[62569]: value = "task-1250224" [ 859.563446] env[62569]: _type = "Task" [ 859.563446] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.573515] env[62569]: DEBUG oslo_vmware.api [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250224, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.574900] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 859.575110] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 859.575850] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f340518d-49e7-45f2-a1e3-506fb8149882 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.581674] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Waiting for the task: (returnval){ [ 859.581674] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52469b8c-979c-ba18-0c10-03040c043691" [ 859.581674] env[62569]: _type = "Task" [ 859.581674] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.589868] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52469b8c-979c-ba18-0c10-03040c043691, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.602197] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250220, 'name': ReconfigVM_Task, 'duration_secs': 0.493611} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.602731] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 31ce29fa-4f60-4404-b830-21ad196f78b5/31ce29fa-4f60-4404-b830-21ad196f78b5.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 859.603108] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-678940fa-91a6-4844-8bda-188b463b3f47 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.610290] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 859.610290] env[62569]: value = "task-1250225" [ 859.610290] env[62569]: _type = "Task" [ 859.610290] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.619453] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250225, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.763364] env[62569]: DEBUG nova.network.neutron [req-9e28a29a-c428-4dd6-96c2-0fa1956c07de req-6fc4f7fc-9b16-4b86-8947-013bf0de5866 service nova] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Updated VIF entry in instance network info cache for port 0edf3387-b21d-47f6-b9e7-503d285ba554. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 859.764294] env[62569]: DEBUG nova.network.neutron [req-9e28a29a-c428-4dd6-96c2-0fa1956c07de req-6fc4f7fc-9b16-4b86-8947-013bf0de5866 service nova] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Updating instance_info_cache with network_info: [{"id": "0edf3387-b21d-47f6-b9e7-503d285ba554", "address": "fa:16:3e:cb:d0:d3", "network": {"id": "e67c1ed0-2a04-4a64-9041-c9dbe857c504", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-566826463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "edea4041206a4fd1bec5b7b62e0922bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24ec44b7-0acf-4ff9-8bb3-4641b74af7a7", "external-id": "nsx-vlan-transportzone-705", "segmentation_id": 705, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0edf3387-b2", "ovs_interfaceid": "0edf3387-b21d-47f6-b9e7-503d285ba554", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.810120] env[62569]: DEBUG oslo_vmware.api [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Task: {'id': task-1250222, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.359685} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.810563] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 859.810867] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 859.811249] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 859.811877] env[62569]: INFO nova.compute.manager [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Took 1.16 seconds to destroy the instance on the hypervisor. [ 859.811974] env[62569]: DEBUG oslo.service.loopingcall [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 859.812250] env[62569]: DEBUG nova.compute.manager [-] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 859.812408] env[62569]: DEBUG nova.network.neutron [-] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 859.881905] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 298efba9-226d-4105-8b32-76b3dddddb0b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 859.882289] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=62569) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 859.882474] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2624MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=62569) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 859.913855] env[62569]: DEBUG nova.network.neutron [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Successfully updated port: 5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 859.935057] env[62569]: DEBUG oslo_concurrency.lockutils [req-9451814b-b513-4af1-94d2-88a8df9a63fc req-f03d80e5-be5c-4fd0-9b22-cf191cd3f33a service nova] Releasing lock "refresh_cache-337d88a2-30b9-4846-929e-042bd7a64a65" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.074594] env[62569]: DEBUG oslo_vmware.api [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250224, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.095611] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52469b8c-979c-ba18-0c10-03040c043691, 'name': SearchDatastore_Task, 'duration_secs': 0.011017} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.096544] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50645d7c-6f0c-4568-84ab-0f1c43e7f886 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.102858] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Waiting for the task: (returnval){ [ 860.102858] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ad81fb-30c2-6f3d-ba9d-3f8471d92f54" [ 860.102858] env[62569]: _type = "Task" [ 860.102858] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.114863] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ad81fb-30c2-6f3d-ba9d-3f8471d92f54, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.123345] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250225, 'name': Rename_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.125091] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f989bdd-b0de-41ad-993f-b2e419b0d027 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.132451] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d259209-b301-4589-8ef1-727986210900 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.168291] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f10e1fd-67a1-45e9-a3d0-2402c9c33a83 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.176850] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c876a7-9992-4c53-9ee3-939348371c5c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.192256] env[62569]: DEBUG nova.compute.provider_tree [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 860.267796] env[62569]: DEBUG oslo_concurrency.lockutils [req-9e28a29a-c428-4dd6-96c2-0fa1956c07de req-6fc4f7fc-9b16-4b86-8947-013bf0de5866 service nova] Releasing lock "refresh_cache-691d8fbe-b9ee-454f-bd7b-14520e53ed26" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.416840] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.417072] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.417268] env[62569]: DEBUG nova.network.neutron [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 860.574626] env[62569]: DEBUG oslo_vmware.api [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250224, 'name': ReconfigVM_Task, 'duration_secs': 0.93307} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.575124] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Reconfigured VM instance instance-00000046 to detach disk 2000 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 860.581654] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62e16057-113a-48dd-98ab-fbf0a8662576 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.600251] env[62569]: DEBUG oslo_vmware.api [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Waiting for the task: (returnval){ [ 860.600251] env[62569]: value = "task-1250226" [ 860.600251] env[62569]: _type = "Task" [ 860.600251] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.615241] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ad81fb-30c2-6f3d-ba9d-3f8471d92f54, 'name': SearchDatastore_Task, 'duration_secs': 0.015597} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.615241] env[62569]: DEBUG oslo_vmware.api [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250226, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.615531] env[62569]: DEBUG oslo_concurrency.lockutils [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.615779] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 691d8fbe-b9ee-454f-bd7b-14520e53ed26/691d8fbe-b9ee-454f-bd7b-14520e53ed26.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 860.618839] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-de162ac1-2903-4262-a3c1-5d3b0fd54c24 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.627483] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250225, 'name': Rename_Task, 'duration_secs': 0.99876} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.629226] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 860.629714] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Waiting for the task: (returnval){ [ 860.629714] env[62569]: value = "task-1250227" [ 860.629714] env[62569]: _type = "Task" [ 860.629714] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.629964] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d6e69078-f6cd-40f2-bdef-01178951bbd2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.642088] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Task: {'id': task-1250227, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.643514] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 860.643514] env[62569]: value = "task-1250228" [ 860.643514] env[62569]: _type = "Task" [ 860.643514] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.651450] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250228, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.669068] env[62569]: DEBUG nova.network.neutron [-] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.718479] env[62569]: ERROR nova.scheduler.client.report [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [req-8b29f39d-58d3-478f-bb31-55f980480508] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID fa06556a-5785-4014-b8bd-bc240a0cf716. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-8b29f39d-58d3-478f-bb31-55f980480508"}]} [ 860.735738] env[62569]: DEBUG nova.scheduler.client.report [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Refreshing inventories for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 860.752087] env[62569]: DEBUG nova.scheduler.client.report [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Updating ProviderTree inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 860.752301] env[62569]: DEBUG nova.compute.provider_tree [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 860.765481] env[62569]: DEBUG nova.scheduler.client.report [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Refreshing aggregate associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, aggregates: None {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 860.786699] env[62569]: DEBUG nova.scheduler.client.report [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Refreshing trait associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 860.982053] env[62569]: WARNING nova.network.neutron [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] 29fcbfe8-35ce-4e8d-874b-2b14ba720adf already exists in list: networks containing: ['29fcbfe8-35ce-4e8d-874b-2b14ba720adf']. ignoring it [ 860.986504] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48610c4c-00ca-4a0c-96a0-9c4b5765434d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.995166] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-929e518d-657a-4428-b44e-357f3f919fbd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.030516] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f78c0947-132d-4c79-996b-8597911c0e2b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.042902] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c91073ff-8265-4ff7-8f04-c0351e4f88f1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.061219] env[62569]: DEBUG nova.compute.provider_tree [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 861.113978] env[62569]: DEBUG oslo_vmware.api [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250226, 'name': ReconfigVM_Task, 'duration_secs': 0.122626} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.114530] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269356', 'volume_id': '3fb6b4c3-cf28-44c4-9606-7430ec94b7f6', 'name': 'volume-3fb6b4c3-cf28-44c4-9606-7430ec94b7f6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'fdeac8b8-a34d-4b4c-8f72-281f84e22ebe', 'attached_at': '', 'detached_at': '', 'volume_id': '3fb6b4c3-cf28-44c4-9606-7430ec94b7f6', 'serial': '3fb6b4c3-cf28-44c4-9606-7430ec94b7f6'} {{(pid=62569) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 861.114648] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 861.115483] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6b051be-4c86-4923-95a9-56cb12762220 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.126646] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 861.126646] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c1abba43-4f25-4cc3-81c2-f288079018b9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.144586] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Task: {'id': task-1250227, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.158247] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250228, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.172319] env[62569]: INFO nova.compute.manager [-] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Took 1.36 seconds to deallocate network for instance. [ 861.194537] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 861.194537] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 861.194537] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Deleting the datastore file [datastore1] fdeac8b8-a34d-4b4c-8f72-281f84e22ebe {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 861.194537] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f12ee824-7c29-4329-80aa-e51ca450e48d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.205873] env[62569]: DEBUG oslo_vmware.api [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Waiting for the task: (returnval){ [ 861.205873] env[62569]: value = "task-1250230" [ 861.205873] env[62569]: _type = "Task" [ 861.205873] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.221635] env[62569]: DEBUG oslo_vmware.api [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250230, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.482687] env[62569]: DEBUG nova.compute.manager [req-b310e4aa-77ff-471e-bfbe-864c1107d81d req-8d83cd07-279f-49c6-b8a6-3ca996044999 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Received event network-vif-plugged-5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 861.482968] env[62569]: DEBUG oslo_concurrency.lockutils [req-b310e4aa-77ff-471e-bfbe-864c1107d81d req-8d83cd07-279f-49c6-b8a6-3ca996044999 service nova] Acquiring lock "807d3025-d6a7-4778-a829-a61e2c7495c9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.483797] env[62569]: DEBUG oslo_concurrency.lockutils [req-b310e4aa-77ff-471e-bfbe-864c1107d81d req-8d83cd07-279f-49c6-b8a6-3ca996044999 service nova] Lock "807d3025-d6a7-4778-a829-a61e2c7495c9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.483797] env[62569]: DEBUG oslo_concurrency.lockutils [req-b310e4aa-77ff-471e-bfbe-864c1107d81d req-8d83cd07-279f-49c6-b8a6-3ca996044999 service nova] Lock "807d3025-d6a7-4778-a829-a61e2c7495c9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.484084] env[62569]: DEBUG nova.compute.manager [req-b310e4aa-77ff-471e-bfbe-864c1107d81d req-8d83cd07-279f-49c6-b8a6-3ca996044999 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] No waiting events found dispatching network-vif-plugged-5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 861.484185] env[62569]: WARNING nova.compute.manager [req-b310e4aa-77ff-471e-bfbe-864c1107d81d req-8d83cd07-279f-49c6-b8a6-3ca996044999 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Received unexpected event network-vif-plugged-5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18 for instance with vm_state active and task_state None. [ 861.484467] env[62569]: DEBUG nova.compute.manager [req-b310e4aa-77ff-471e-bfbe-864c1107d81d req-8d83cd07-279f-49c6-b8a6-3ca996044999 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Received event network-changed-5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 861.484587] env[62569]: DEBUG nova.compute.manager [req-b310e4aa-77ff-471e-bfbe-864c1107d81d req-8d83cd07-279f-49c6-b8a6-3ca996044999 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Refreshing instance network info cache due to event network-changed-5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 861.484704] env[62569]: DEBUG oslo_concurrency.lockutils [req-b310e4aa-77ff-471e-bfbe-864c1107d81d req-8d83cd07-279f-49c6-b8a6-3ca996044999 service nova] Acquiring lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.485786] env[62569]: DEBUG nova.network.neutron [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Updating instance_info_cache with network_info: [{"id": "9d855102-33ab-4574-b791-dfa6fb5b0f09", "address": "fa:16:3e:b3:4b:f5", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d855102-33", "ovs_interfaceid": "9d855102-33ab-4574-b791-dfa6fb5b0f09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18", "address": "fa:16:3e:df:90:ac", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f2ad2ec-3e", "ovs_interfaceid": "5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.500271] env[62569]: DEBUG nova.compute.manager [req-d51bbe18-69fb-4342-800a-34bc6366f156 req-291ea594-880e-412a-b907-8d707815d25c service nova] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Received event network-vif-deleted-96148207-1e82-4a26-baa5-65f88f889f97 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 861.596298] env[62569]: DEBUG nova.scheduler.client.report [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Updated inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with generation 91 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 861.596600] env[62569]: DEBUG nova.compute.provider_tree [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Updating resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 generation from 91 to 92 during operation: update_inventory {{(pid=62569) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 861.596776] env[62569]: DEBUG nova.compute.provider_tree [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 861.644149] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Task: {'id': task-1250227, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.531023} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.644537] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 691d8fbe-b9ee-454f-bd7b-14520e53ed26/691d8fbe-b9ee-454f-bd7b-14520e53ed26.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 861.644757] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 861.648659] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-38274b95-8be3-432d-83b9-226bd6964e9f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.659905] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Waiting for the task: (returnval){ [ 861.659905] env[62569]: value = "task-1250231" [ 861.659905] env[62569]: _type = "Task" [ 861.659905] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.660166] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250228, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.668956] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Task: {'id': task-1250231, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.680768] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.717986] env[62569]: DEBUG oslo_vmware.api [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Task: {'id': task-1250230, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159032} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.719110] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 861.719391] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 861.719688] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 861.786599] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Volume detach. Driver type: vmdk {{(pid=62569) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 861.787060] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-04b2d125-6cab-4a8e-905b-7340c8b2faef {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.799022] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92e6f5ff-79c8-4374-b844-913d6c1f8b8e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.831723] env[62569]: ERROR nova.compute.manager [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Failed to detach volume 3fb6b4c3-cf28-44c4-9606-7430ec94b7f6 from /dev/sda: nova.exception.InstanceNotFound: Instance fdeac8b8-a34d-4b4c-8f72-281f84e22ebe could not be found. [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Traceback (most recent call last): [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/compute/manager.py", line 4184, in _do_rebuild_instance [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] self.driver.rebuild(**kwargs) [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/virt/driver.py", line 497, in rebuild [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] raise NotImplementedError() [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] NotImplementedError [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] During handling of the above exception, another exception occurred: [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Traceback (most recent call last): [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/compute/manager.py", line 3607, in _detach_root_volume [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] self.driver.detach_volume(context, old_connection_info, [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 555, in detach_volume [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] return self._volumeops.detach_volume(connection_info, instance) [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] self._detach_volume_vmdk(connection_info, instance) [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] stable_ref.fetch_moref(session) [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] raise exception.InstanceNotFound(instance_id=self._uuid) [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] nova.exception.InstanceNotFound: Instance fdeac8b8-a34d-4b4c-8f72-281f84e22ebe could not be found. [ 861.831723] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] [ 861.968831] env[62569]: DEBUG nova.compute.utils [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Build of instance fdeac8b8-a34d-4b4c-8f72-281f84e22ebe aborted: Failed to rebuild volume backed instance. {{(pid=62569) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 861.971513] env[62569]: ERROR nova.compute.manager [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance fdeac8b8-a34d-4b4c-8f72-281f84e22ebe aborted: Failed to rebuild volume backed instance. [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Traceback (most recent call last): [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/compute/manager.py", line 4184, in _do_rebuild_instance [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] self.driver.rebuild(**kwargs) [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/virt/driver.py", line 497, in rebuild [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] raise NotImplementedError() [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] NotImplementedError [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] During handling of the above exception, another exception occurred: [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Traceback (most recent call last): [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/compute/manager.py", line 3642, in _rebuild_volume_backed_instance [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] self._detach_root_volume(context, instance, root_bdm) [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/compute/manager.py", line 3621, in _detach_root_volume [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] with excutils.save_and_reraise_exception(): [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] self.force_reraise() [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] raise self.value [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/compute/manager.py", line 3607, in _detach_root_volume [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] self.driver.detach_volume(context, old_connection_info, [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 555, in detach_volume [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] return self._volumeops.detach_volume(connection_info, instance) [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] self._detach_volume_vmdk(connection_info, instance) [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] stable_ref.fetch_moref(session) [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] raise exception.InstanceNotFound(instance_id=self._uuid) [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] nova.exception.InstanceNotFound: Instance fdeac8b8-a34d-4b4c-8f72-281f84e22ebe could not be found. [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] During handling of the above exception, another exception occurred: [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Traceback (most recent call last): [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/compute/manager.py", line 11245, in _error_out_instance_on_exception [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] yield [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/compute/manager.py", line 3910, in rebuild_instance [ 861.971513] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] self._do_rebuild_instance_with_claim( [ 861.972942] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/compute/manager.py", line 3996, in _do_rebuild_instance_with_claim [ 861.972942] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] self._do_rebuild_instance( [ 861.972942] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/compute/manager.py", line 4188, in _do_rebuild_instance [ 861.972942] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] self._rebuild_default_impl(**kwargs) [ 861.972942] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/compute/manager.py", line 3765, in _rebuild_default_impl [ 861.972942] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] self._rebuild_volume_backed_instance( [ 861.972942] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] File "/opt/stack/nova/nova/compute/manager.py", line 3657, in _rebuild_volume_backed_instance [ 861.972942] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] raise exception.BuildAbortException( [ 861.972942] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] nova.exception.BuildAbortException: Build of instance fdeac8b8-a34d-4b4c-8f72-281f84e22ebe aborted: Failed to rebuild volume backed instance. [ 861.972942] env[62569]: ERROR nova.compute.manager [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] [ 861.988462] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.989213] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.989750] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.989750] env[62569]: DEBUG oslo_concurrency.lockutils [req-b310e4aa-77ff-471e-bfbe-864c1107d81d req-8d83cd07-279f-49c6-b8a6-3ca996044999 service nova] Acquired lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.989886] env[62569]: DEBUG nova.network.neutron [req-b310e4aa-77ff-471e-bfbe-864c1107d81d req-8d83cd07-279f-49c6-b8a6-3ca996044999 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Refreshing network info cache for port 5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 861.992265] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d84f54ec-d894-4275-8807-fa0bc78581a2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.010310] env[62569]: DEBUG nova.virt.hardware [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 862.010577] env[62569]: DEBUG nova.virt.hardware [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 862.010743] env[62569]: DEBUG nova.virt.hardware [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 862.010937] env[62569]: DEBUG nova.virt.hardware [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 862.011107] env[62569]: DEBUG nova.virt.hardware [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 862.011263] env[62569]: DEBUG nova.virt.hardware [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 862.012099] env[62569]: DEBUG nova.virt.hardware [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 862.012099] env[62569]: DEBUG nova.virt.hardware [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 862.012099] env[62569]: DEBUG nova.virt.hardware [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 862.012099] env[62569]: DEBUG nova.virt.hardware [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 862.012245] env[62569]: DEBUG nova.virt.hardware [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 862.018979] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Reconfiguring VM to attach interface {{(pid=62569) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 862.020907] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f035a812-a441-4e76-83e6-54e4121f8bbb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.039590] env[62569]: DEBUG oslo_vmware.api [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 862.039590] env[62569]: value = "task-1250232" [ 862.039590] env[62569]: _type = "Task" [ 862.039590] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.047568] env[62569]: DEBUG oslo_vmware.api [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250232, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.105698] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62569) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 862.105975] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.776s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.106505] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.401s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.106943] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.109309] env[62569]: DEBUG oslo_concurrency.lockutils [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.588s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.109729] env[62569]: DEBUG oslo_concurrency.lockutils [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.113054] env[62569]: DEBUG oslo_concurrency.lockutils [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.926s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.114898] env[62569]: INFO nova.compute.claims [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 862.143240] env[62569]: INFO nova.scheduler.client.report [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Deleted allocations for instance c8edf395-ea21-4ba8-991e-0f43fe4ee830 [ 862.160643] env[62569]: DEBUG oslo_vmware.api [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250228, 'name': PowerOnVM_Task, 'duration_secs': 1.088436} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.161065] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 862.161292] env[62569]: INFO nova.compute.manager [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Took 10.43 seconds to spawn the instance on the hypervisor. [ 862.161477] env[62569]: DEBUG nova.compute.manager [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 862.165868] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b936d66-185f-4cff-8f6a-839f456f5699 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.175927] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Task: {'id': task-1250231, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075832} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.177673] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 862.182763] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a88cadc-15a4-48ec-ba99-79c44b314f69 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.213852] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] 691d8fbe-b9ee-454f-bd7b-14520e53ed26/691d8fbe-b9ee-454f-bd7b-14520e53ed26.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 862.214943] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df5f0ff0-c3eb-49ae-abd6-d684c02bfb6a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.236370] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Waiting for the task: (returnval){ [ 862.236370] env[62569]: value = "task-1250233" [ 862.236370] env[62569]: _type = "Task" [ 862.236370] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.244717] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Task: {'id': task-1250233, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.549977] env[62569]: DEBUG oslo_vmware.api [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250232, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.619677] env[62569]: DEBUG oslo_concurrency.lockutils [None req-89c6f0bc-2688-4dbe-b2bb-67d3e52f0204 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "337d88a2-30b9-4846-929e-042bd7a64a65" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 27.254s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.651554] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6b319570-e668-4e27-be47-3c80d7014752 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "c8edf395-ea21-4ba8-991e-0f43fe4ee830" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.413s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.691694] env[62569]: DEBUG oslo_concurrency.lockutils [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquiring lock "337d88a2-30b9-4846-929e-042bd7a64a65" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.691967] env[62569]: DEBUG oslo_concurrency.lockutils [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "337d88a2-30b9-4846-929e-042bd7a64a65" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.692179] env[62569]: INFO nova.compute.manager [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Unshelving [ 862.706290] env[62569]: INFO nova.compute.manager [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Took 44.21 seconds to build instance. [ 862.749377] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Task: {'id': task-1250233, 'name': ReconfigVM_Task, 'duration_secs': 0.326753} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.749727] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Reconfigured VM instance instance-00000050 to attach disk [datastore1] 691d8fbe-b9ee-454f-bd7b-14520e53ed26/691d8fbe-b9ee-454f-bd7b-14520e53ed26.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 862.750446] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-02309856-3bd4-41dc-b928-7d7a541a3ecf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.757502] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Waiting for the task: (returnval){ [ 862.757502] env[62569]: value = "task-1250234" [ 862.757502] env[62569]: _type = "Task" [ 862.757502] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.765968] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Task: {'id': task-1250234, 'name': Rename_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.842777] env[62569]: DEBUG nova.network.neutron [req-b310e4aa-77ff-471e-bfbe-864c1107d81d req-8d83cd07-279f-49c6-b8a6-3ca996044999 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Updated VIF entry in instance network info cache for port 5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 862.843437] env[62569]: DEBUG nova.network.neutron [req-b310e4aa-77ff-471e-bfbe-864c1107d81d req-8d83cd07-279f-49c6-b8a6-3ca996044999 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Updating instance_info_cache with network_info: [{"id": "9d855102-33ab-4574-b791-dfa6fb5b0f09", "address": "fa:16:3e:b3:4b:f5", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d855102-33", "ovs_interfaceid": "9d855102-33ab-4574-b791-dfa6fb5b0f09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18", "address": "fa:16:3e:df:90:ac", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f2ad2ec-3e", "ovs_interfaceid": "5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.935812] env[62569]: DEBUG nova.compute.manager [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 862.937120] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b52367a6-1b5b-4476-bb90-03ef7a865a30 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.050129] env[62569]: DEBUG oslo_vmware.api [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250232, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.208479] env[62569]: DEBUG oslo_concurrency.lockutils [None req-399dc830-71e3-453f-b56e-6ca67db693f3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "31ce29fa-4f60-4404-b830-21ad196f78b5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.954s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.270229] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Task: {'id': task-1250234, 'name': Rename_Task, 'duration_secs': 0.156515} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.270492] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 863.270751] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7d51785a-a062-4c93-bde9-7c56e0651d79 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.278173] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Waiting for the task: (returnval){ [ 863.278173] env[62569]: value = "task-1250235" [ 863.278173] env[62569]: _type = "Task" [ 863.278173] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.286731] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Task: {'id': task-1250235, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.313281] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae1bc630-dc60-42ed-adbc-aa271b33a39f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.321470] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc2f04dd-c28e-4116-9d12-f301d3b7cfdc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.351768] env[62569]: DEBUG oslo_concurrency.lockutils [req-b310e4aa-77ff-471e-bfbe-864c1107d81d req-8d83cd07-279f-49c6-b8a6-3ca996044999 service nova] Releasing lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.353029] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9004f637-faf1-41bd-b47d-d849dd309182 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.360740] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-539aa758-8621-447d-8b82-8e652fb347ed {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.375974] env[62569]: DEBUG nova.compute.provider_tree [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.450287] env[62569]: INFO nova.compute.manager [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] instance snapshotting [ 863.451078] env[62569]: DEBUG nova.objects.instance [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lazy-loading 'flavor' on Instance uuid b47cd2d7-0cd2-41af-8ed1-a6dfca323516 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 863.552159] env[62569]: DEBUG oslo_vmware.api [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250232, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.729213] env[62569]: DEBUG oslo_concurrency.lockutils [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.790246] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Task: {'id': task-1250235, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.839040] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "108d5bde-e463-4a69-816b-bdd3c03e13cd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.839040] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "108d5bde-e463-4a69-816b-bdd3c03e13cd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.879208] env[62569]: DEBUG nova.scheduler.client.report [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 863.960914] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a9a614-9cd3-4ebe-8647-83f1ef00f0f2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.983397] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ae1593f-cbb6-4679-840d-71b7b0629924 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.987275] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.052975] env[62569]: DEBUG oslo_vmware.api [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250232, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.292971] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Task: {'id': task-1250235, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.347828] env[62569]: DEBUG nova.compute.manager [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 864.388404] env[62569]: DEBUG oslo_concurrency.lockutils [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.275s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.389107] env[62569]: DEBUG nova.compute.manager [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 864.393859] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.713s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.394016] env[62569]: DEBUG nova.objects.instance [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Lazy-loading 'resources' on Instance uuid a5b3c345-149b-4dc3-84ff-fe72183efadb {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 864.497214] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Creating Snapshot of the VM instance {{(pid=62569) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 864.497214] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-6b83a01b-3d27-4891-9d3d-29f0b44312b3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.506055] env[62569]: DEBUG oslo_vmware.api [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 864.506055] env[62569]: value = "task-1250236" [ 864.506055] env[62569]: _type = "Task" [ 864.506055] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.521137] env[62569]: DEBUG oslo_vmware.api [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250236, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.553669] env[62569]: DEBUG oslo_vmware.api [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250232, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.796510] env[62569]: DEBUG oslo_vmware.api [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Task: {'id': task-1250235, 'name': PowerOnVM_Task, 'duration_secs': 1.265312} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.796510] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 864.796510] env[62569]: INFO nova.compute.manager [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Took 8.76 seconds to spawn the instance on the hypervisor. [ 864.796510] env[62569]: DEBUG nova.compute.manager [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 864.796510] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e931c03-c4b8-4f8d-b1e9-56a789b4d357 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.882244] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.901222] env[62569]: DEBUG nova.compute.utils [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 864.905151] env[62569]: DEBUG nova.compute.manager [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 864.905462] env[62569]: DEBUG nova.network.neutron [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 865.006680] env[62569]: DEBUG nova.policy [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5eab0a9d024d4d7e9fa6ebe42c88db12', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1af23d0979b84b85b348b3868379ef2b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 865.025276] env[62569]: DEBUG oslo_vmware.api [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250236, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.027669] env[62569]: DEBUG oslo_vmware.rw_handles [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5237f662-615a-a9d3-b3df-ef2a6b53a3ea/disk-0.vmdk. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 865.029104] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97bbd1ca-7466-4f5e-8166-d422c9f81a31 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.038651] env[62569]: DEBUG oslo_vmware.rw_handles [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5237f662-615a-a9d3-b3df-ef2a6b53a3ea/disk-0.vmdk is in state: ready. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 865.038942] env[62569]: ERROR oslo_vmware.rw_handles [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5237f662-615a-a9d3-b3df-ef2a6b53a3ea/disk-0.vmdk due to incomplete transfer. [ 865.039247] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-8d8670ca-cb3a-4616-a408-cd305dac3ee4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.049659] env[62569]: DEBUG oslo_vmware.rw_handles [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5237f662-615a-a9d3-b3df-ef2a6b53a3ea/disk-0.vmdk. {{(pid=62569) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 865.050178] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Uploaded image 5909e6a6-6c00-4956-a4d0-c6430d91b165 to the Glance image server {{(pid=62569) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 865.052586] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Destroying the VM {{(pid=62569) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 865.056192] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4f6048e9-d93c-4ce0-bef7-1368f3a6045e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.061336] env[62569]: DEBUG oslo_vmware.api [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250232, 'name': ReconfigVM_Task, 'duration_secs': 2.712501} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.062250] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.062584] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Reconfigured VM to attach interface {{(pid=62569) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 865.069091] env[62569]: DEBUG oslo_vmware.api [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 865.069091] env[62569]: value = "task-1250237" [ 865.069091] env[62569]: _type = "Task" [ 865.069091] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.080668] env[62569]: DEBUG oslo_vmware.api [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250237, 'name': Destroy_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.142381] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af347a0-34a9-49a7-9aa3-e4d04770f02d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.150528] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-591bd107-ca7e-41b0-afb5-931e39bb6778 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.186832] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bacafe90-c79a-40f5-a0a5-63eb6356fe0e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.194858] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be37eba9-14c7-4083-aad0-29e7741f3ff1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.209061] env[62569]: DEBUG nova.compute.provider_tree [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 865.320235] env[62569]: INFO nova.compute.manager [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Took 30.81 seconds to build instance. [ 865.342974] env[62569]: DEBUG oslo_concurrency.lockutils [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Acquiring lock "fdeac8b8-a34d-4b4c-8f72-281f84e22ebe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.343217] env[62569]: DEBUG oslo_concurrency.lockutils [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Lock "fdeac8b8-a34d-4b4c-8f72-281f84e22ebe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.343583] env[62569]: DEBUG oslo_concurrency.lockutils [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Acquiring lock "fdeac8b8-a34d-4b4c-8f72-281f84e22ebe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.344637] env[62569]: DEBUG oslo_concurrency.lockutils [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Lock "fdeac8b8-a34d-4b4c-8f72-281f84e22ebe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.344831] env[62569]: DEBUG oslo_concurrency.lockutils [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Lock "fdeac8b8-a34d-4b4c-8f72-281f84e22ebe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.350037] env[62569]: INFO nova.compute.manager [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Terminating instance [ 865.392154] env[62569]: DEBUG nova.network.neutron [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Successfully created port: f33064d8-b284-4ed3-9a2b-01f27371be83 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 865.408582] env[62569]: DEBUG nova.compute.manager [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 865.523713] env[62569]: DEBUG oslo_vmware.api [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250236, 'name': CreateSnapshot_Task, 'duration_secs': 0.884864} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.524783] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Created Snapshot of the VM instance {{(pid=62569) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 865.524931] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a01bd517-c6b3-41eb-9200-b693c153ad80 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.568506] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0d7b6e25-7adb-4ac3-9156-8522caae0a98 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "interface-807d3025-d6a7-4778-a829-a61e2c7495c9-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 9.439s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.578667] env[62569]: DEBUG oslo_vmware.api [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250237, 'name': Destroy_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.714219] env[62569]: DEBUG nova.scheduler.client.report [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 865.822140] env[62569]: DEBUG oslo_concurrency.lockutils [None req-baf1c833-97c5-450d-a9bc-b77be3c05f01 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Lock "691d8fbe-b9ee-454f-bd7b-14520e53ed26" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.325s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.854427] env[62569]: DEBUG nova.compute.manager [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 865.855267] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dbc4290e-eeae-4fc8-b3ac-ba1e895067da {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.865938] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd29e12-0025-4a8e-8b89-5568debfc206 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.881206] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f1f3a574-b618-4a13-86f3-686b2e7db7f7 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "cb9d8991-b7ba-4bf7-9d22-1391f4cea768" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.881448] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f1f3a574-b618-4a13-86f3-686b2e7db7f7 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cb9d8991-b7ba-4bf7-9d22-1391f4cea768" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.896197] env[62569]: WARNING nova.virt.vmwareapi.driver [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance fdeac8b8-a34d-4b4c-8f72-281f84e22ebe could not be found. [ 865.896411] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 865.897229] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e335247d-5d6c-4cd8-bbaf-fd0b1497b3ea {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.905981] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd504b7b-2570-4814-9867-0a50f1d36035 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.939302] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fdeac8b8-a34d-4b4c-8f72-281f84e22ebe could not be found. [ 865.939596] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 865.939795] env[62569]: INFO nova.compute.manager [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Took 0.09 seconds to destroy the instance on the hypervisor. [ 865.940420] env[62569]: DEBUG oslo.service.loopingcall [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 865.940785] env[62569]: DEBUG nova.compute.manager [-] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 865.940785] env[62569]: DEBUG nova.network.neutron [-] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 866.046211] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Creating linked-clone VM from snapshot {{(pid=62569) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 866.046627] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8ad568f8-71e1-43eb-8f23-f88b50fb9125 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.055013] env[62569]: DEBUG oslo_vmware.api [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 866.055013] env[62569]: value = "task-1250238" [ 866.055013] env[62569]: _type = "Task" [ 866.055013] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.063686] env[62569]: DEBUG oslo_vmware.api [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250238, 'name': CloneVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.077894] env[62569]: DEBUG oslo_vmware.api [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250237, 'name': Destroy_Task, 'duration_secs': 0.769413} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.078182] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Destroyed the VM [ 866.078426] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Deleting Snapshot of the VM instance {{(pid=62569) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 866.078677] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-76002767-a4d8-4324-b57a-495cb4afa380 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.085261] env[62569]: DEBUG oslo_vmware.api [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 866.085261] env[62569]: value = "task-1250239" [ 866.085261] env[62569]: _type = "Task" [ 866.085261] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.094472] env[62569]: DEBUG oslo_vmware.api [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250239, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.220134] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.827s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.224211] env[62569]: DEBUG oslo_concurrency.lockutils [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.495s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.224474] env[62569]: DEBUG nova.objects.instance [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lazy-loading 'pci_requests' on Instance uuid 337d88a2-30b9-4846-929e-042bd7a64a65 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 866.251456] env[62569]: INFO nova.scheduler.client.report [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Deleted allocations for instance a5b3c345-149b-4dc3-84ff-fe72183efadb [ 866.384931] env[62569]: DEBUG nova.compute.utils [None req-f1f3a574-b618-4a13-86f3-686b2e7db7f7 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 866.421996] env[62569]: DEBUG nova.compute.manager [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 866.450683] env[62569]: DEBUG nova.virt.hardware [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 866.451030] env[62569]: DEBUG nova.virt.hardware [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 866.451337] env[62569]: DEBUG nova.virt.hardware [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 866.451589] env[62569]: DEBUG nova.virt.hardware [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 866.451786] env[62569]: DEBUG nova.virt.hardware [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 866.452012] env[62569]: DEBUG nova.virt.hardware [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 866.452299] env[62569]: DEBUG nova.virt.hardware [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 866.452508] env[62569]: DEBUG nova.virt.hardware [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 866.452781] env[62569]: DEBUG nova.virt.hardware [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 866.452941] env[62569]: DEBUG nova.virt.hardware [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 866.453286] env[62569]: DEBUG nova.virt.hardware [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 866.455576] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-330b2871-db7d-4a12-b860-d749bbca874f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.463733] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbef87ef-93f5-49ba-b931-4384f44c34c4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.566411] env[62569]: DEBUG oslo_vmware.api [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250238, 'name': CloneVM_Task} progress is 94%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.596752] env[62569]: DEBUG oslo_vmware.api [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250239, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.729232] env[62569]: DEBUG nova.objects.instance [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lazy-loading 'numa_topology' on Instance uuid 337d88a2-30b9-4846-929e-042bd7a64a65 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 866.763434] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1ae52499-8f67-4428-a7fb-cc23f23f9bdc tempest-ServerRescueTestJSONUnderV235-530616678 tempest-ServerRescueTestJSONUnderV235-530616678-project-member] Lock "a5b3c345-149b-4dc3-84ff-fe72183efadb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.625s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.889907] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f1f3a574-b618-4a13-86f3-686b2e7db7f7 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cb9d8991-b7ba-4bf7-9d22-1391f4cea768" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.960693] env[62569]: DEBUG nova.network.neutron [-] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.073457] env[62569]: DEBUG oslo_vmware.api [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250238, 'name': CloneVM_Task} progress is 94%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.103071] env[62569]: DEBUG oslo_vmware.api [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250239, 'name': RemoveSnapshot_Task, 'duration_secs': 0.722132} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.103525] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Deleted Snapshot of the VM instance {{(pid=62569) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 867.103900] env[62569]: INFO nova.compute.manager [None req-37ca5d2e-c57d-4c42-8d6a-72f8e12766ca tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Took 15.51 seconds to snapshot the instance on the hypervisor. [ 867.232105] env[62569]: DEBUG nova.compute.manager [req-e8c084f1-45ea-44ce-bba6-09fcfc08c85c req-5f692230-7ca4-4ceb-b2c5-7acdbeebbbaf service nova] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Received event network-vif-deleted-da0de7a0-922e-46b2-842b-a222e31de3b2 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 867.234116] env[62569]: INFO nova.compute.claims [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 867.458271] env[62569]: DEBUG nova.network.neutron [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Successfully updated port: f33064d8-b284-4ed3-9a2b-01f27371be83 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 867.465921] env[62569]: INFO nova.compute.manager [-] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Took 1.52 seconds to deallocate network for instance. [ 867.568499] env[62569]: DEBUG oslo_vmware.api [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250238, 'name': CloneVM_Task} progress is 95%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.749854] env[62569]: DEBUG nova.compute.manager [req-d734f935-69db-471e-8eb7-914cb1e78614 req-9856af44-9dbd-4d8e-88f4-d951cdd37fae service nova] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Received event network-vif-plugged-f33064d8-b284-4ed3-9a2b-01f27371be83 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 867.749854] env[62569]: DEBUG oslo_concurrency.lockutils [req-d734f935-69db-471e-8eb7-914cb1e78614 req-9856af44-9dbd-4d8e-88f4-d951cdd37fae service nova] Acquiring lock "298efba9-226d-4105-8b32-76b3dddddb0b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.749854] env[62569]: DEBUG oslo_concurrency.lockutils [req-d734f935-69db-471e-8eb7-914cb1e78614 req-9856af44-9dbd-4d8e-88f4-d951cdd37fae service nova] Lock "298efba9-226d-4105-8b32-76b3dddddb0b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.749854] env[62569]: DEBUG oslo_concurrency.lockutils [req-d734f935-69db-471e-8eb7-914cb1e78614 req-9856af44-9dbd-4d8e-88f4-d951cdd37fae service nova] Lock "298efba9-226d-4105-8b32-76b3dddddb0b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.749854] env[62569]: DEBUG nova.compute.manager [req-d734f935-69db-471e-8eb7-914cb1e78614 req-9856af44-9dbd-4d8e-88f4-d951cdd37fae service nova] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] No waiting events found dispatching network-vif-plugged-f33064d8-b284-4ed3-9a2b-01f27371be83 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 867.751298] env[62569]: WARNING nova.compute.manager [req-d734f935-69db-471e-8eb7-914cb1e78614 req-9856af44-9dbd-4d8e-88f4-d951cdd37fae service nova] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Received unexpected event network-vif-plugged-f33064d8-b284-4ed3-9a2b-01f27371be83 for instance with vm_state building and task_state spawning. [ 867.967592] env[62569]: DEBUG oslo_concurrency.lockutils [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Acquiring lock "refresh_cache-298efba9-226d-4105-8b32-76b3dddddb0b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.967592] env[62569]: DEBUG oslo_concurrency.lockutils [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Acquired lock "refresh_cache-298efba9-226d-4105-8b32-76b3dddddb0b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.967592] env[62569]: DEBUG nova.network.neutron [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 867.975423] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-136ffef9-ad26-4d84-a8f7-9e02352629da {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.982194] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f1f3a574-b618-4a13-86f3-686b2e7db7f7 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "cb9d8991-b7ba-4bf7-9d22-1391f4cea768" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.982472] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f1f3a574-b618-4a13-86f3-686b2e7db7f7 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cb9d8991-b7ba-4bf7-9d22-1391f4cea768" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.982915] env[62569]: INFO nova.compute.manager [None req-f1f3a574-b618-4a13-86f3-686b2e7db7f7 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Attaching volume f29f7945-6df2-4fc4-bd59-5d5daee539e7 to /dev/sdb [ 867.988831] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3554e3f6-4493-4713-8acd-6fec1ffb1473 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.033275] env[62569]: INFO nova.compute.manager [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Took 0.57 seconds to detach 1 volumes for instance. [ 868.035518] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1c3e7b3-e4be-4f62-9eb5-1373f5f44c5b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.039216] env[62569]: DEBUG nova.compute.manager [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Deleting volume: 3fb6b4c3-cf28-44c4-9606-7430ec94b7f6 {{(pid=62569) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3281}} [ 868.041454] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fefb24ab-fb25-4bda-9045-28472b7feb82 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.053460] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a9d8d51-3e9c-41ce-bc23-7d3efd453b96 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.060046] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40634334-abc9-4b7f-a757-b8fd1fa2654e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.075646] env[62569]: DEBUG nova.compute.provider_tree [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.083137] env[62569]: DEBUG oslo_vmware.api [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250238, 'name': CloneVM_Task, 'duration_secs': 1.891201} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.083797] env[62569]: DEBUG nova.virt.block_device [None req-f1f3a574-b618-4a13-86f3-686b2e7db7f7 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Updating existing volume attachment record: 92e9990a-9aa3-4613-bfa9-85f33a2f2dca {{(pid=62569) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 868.094315] env[62569]: DEBUG nova.scheduler.client.report [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 868.097882] env[62569]: INFO nova.virt.vmwareapi.vmops [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Created linked-clone VM from snapshot [ 868.099226] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a079b896-7260-4c0d-ae45-1b97f57a3475 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.109286] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Uploading image 870dec0b-8c00-43e0-870e-316f6bd64e1e {{(pid=62569) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 868.147795] env[62569]: DEBUG oslo_vmware.rw_handles [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 868.147795] env[62569]: value = "vm-269444" [ 868.147795] env[62569]: _type = "VirtualMachine" [ 868.147795] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 868.148739] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ca647b40-a068-4402-b913-ba55e614030c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.155736] env[62569]: DEBUG oslo_vmware.rw_handles [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lease: (returnval){ [ 868.155736] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]525dca16-101f-6dc0-7b2d-61f3d13e260f" [ 868.155736] env[62569]: _type = "HttpNfcLease" [ 868.155736] env[62569]: } obtained for exporting VM: (result){ [ 868.155736] env[62569]: value = "vm-269444" [ 868.155736] env[62569]: _type = "VirtualMachine" [ 868.155736] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 868.156543] env[62569]: DEBUG oslo_vmware.api [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the lease: (returnval){ [ 868.156543] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]525dca16-101f-6dc0-7b2d-61f3d13e260f" [ 868.156543] env[62569]: _type = "HttpNfcLease" [ 868.156543] env[62569]: } to be ready. {{(pid=62569) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 868.166433] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 868.166433] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]525dca16-101f-6dc0-7b2d-61f3d13e260f" [ 868.166433] env[62569]: _type = "HttpNfcLease" [ 868.166433] env[62569]: } is initializing. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 868.313160] env[62569]: DEBUG nova.compute.manager [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Stashing vm_state: active {{(pid=62569) _prep_resize /opt/stack/nova/nova/compute/manager.py:5997}} [ 868.540537] env[62569]: DEBUG nova.network.neutron [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 868.603159] env[62569]: DEBUG oslo_concurrency.lockutils [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.604025] env[62569]: DEBUG oslo_concurrency.lockutils [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.380s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.606700] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.619s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.669652] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 868.669652] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]525dca16-101f-6dc0-7b2d-61f3d13e260f" [ 868.669652] env[62569]: _type = "HttpNfcLease" [ 868.669652] env[62569]: } is ready. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 868.669652] env[62569]: DEBUG oslo_vmware.rw_handles [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 868.669652] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]525dca16-101f-6dc0-7b2d-61f3d13e260f" [ 868.669652] env[62569]: _type = "HttpNfcLease" [ 868.669652] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 868.670411] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59c6d983-8197-4b93-a4ad-55fb257d7b0a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.680605] env[62569]: DEBUG oslo_vmware.rw_handles [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527173ad-8bf5-4446-8686-6db5d632a24f/disk-0.vmdk from lease info. {{(pid=62569) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 868.683815] env[62569]: DEBUG oslo_vmware.rw_handles [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527173ad-8bf5-4446-8686-6db5d632a24f/disk-0.vmdk for reading. {{(pid=62569) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 868.758668] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "interface-807d3025-d6a7-4778-a829-a61e2c7495c9-62c1052e-dc0a-4fb6-a399-3e238bf93789" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.758926] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "interface-807d3025-d6a7-4778-a829-a61e2c7495c9-62c1052e-dc0a-4fb6-a399-3e238bf93789" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.759363] env[62569]: DEBUG nova.objects.instance [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lazy-loading 'flavor' on Instance uuid 807d3025-d6a7-4778-a829-a61e2c7495c9 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 868.794765] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2381d930-0b27-420e-8bad-4390ea8fcf94 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.839776] env[62569]: DEBUG oslo_concurrency.lockutils [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.888189] env[62569]: INFO nova.network.neutron [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Updating port 303f46d4-6e0a-418c-b62b-17323ab645b6 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 868.958908] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-451e3808-4f29-43fb-9f90-3058943a5d2d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.969234] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-990daa5a-305b-4cda-8960-b4f596fc8985 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.022237] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60052d72-1894-4136-abe8-e60bcb2bc64e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.031760] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-533aa6ed-1d2f-4195-8249-3765becc3c07 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.052845] env[62569]: DEBUG nova.compute.provider_tree [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 869.170813] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "9c1c0379-3968-464b-b587-8fac704404d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.171286] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "9c1c0379-3968-464b-b587-8fac704404d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.238181] env[62569]: DEBUG nova.network.neutron [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Updating instance_info_cache with network_info: [{"id": "f33064d8-b284-4ed3-9a2b-01f27371be83", "address": "fa:16:3e:0b:05:c4", "network": {"id": "8a82d478-4aa8-4179-b974-51200b4e2bc6", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1314992403-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1af23d0979b84b85b348b3868379ef2b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf33064d8-b2", "ovs_interfaceid": "f33064d8-b284-4ed3-9a2b-01f27371be83", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.521790] env[62569]: DEBUG nova.objects.instance [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lazy-loading 'pci_requests' on Instance uuid 807d3025-d6a7-4778-a829-a61e2c7495c9 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 869.556202] env[62569]: DEBUG nova.scheduler.client.report [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 869.673724] env[62569]: DEBUG nova.compute.manager [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 869.741649] env[62569]: DEBUG oslo_concurrency.lockutils [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Releasing lock "refresh_cache-298efba9-226d-4105-8b32-76b3dddddb0b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.742151] env[62569]: DEBUG nova.compute.manager [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Instance network_info: |[{"id": "f33064d8-b284-4ed3-9a2b-01f27371be83", "address": "fa:16:3e:0b:05:c4", "network": {"id": "8a82d478-4aa8-4179-b974-51200b4e2bc6", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1314992403-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1af23d0979b84b85b348b3868379ef2b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf33064d8-b2", "ovs_interfaceid": "f33064d8-b284-4ed3-9a2b-01f27371be83", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 869.743928] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:05:c4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e445fb59-822c-4d7d-943b-c8e3bbaca62e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f33064d8-b284-4ed3-9a2b-01f27371be83', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 869.757759] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Creating folder: Project (1af23d0979b84b85b348b3868379ef2b). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 869.759082] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-556df523-0e6e-455c-9da3-8508756a03ba {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.772599] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Created folder: Project (1af23d0979b84b85b348b3868379ef2b) in parent group-v269330. [ 869.772971] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Creating folder: Instances. Parent ref: group-v269447. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 869.773308] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-116e99c4-6af8-4395-8a23-15bd34f6381e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.787175] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Created folder: Instances in parent group-v269447. [ 869.787175] env[62569]: DEBUG oslo.service.loopingcall [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 869.787420] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 869.787707] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-57fecfbe-8f67-4176-9d84-165cd4d30b43 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.818992] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 869.818992] env[62569]: value = "task-1250247" [ 869.818992] env[62569]: _type = "Task" [ 869.818992] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.827570] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250247, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.025822] env[62569]: DEBUG nova.objects.base [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Object Instance<807d3025-d6a7-4778-a829-a61e2c7495c9> lazy-loaded attributes: flavor,pci_requests {{(pid=62569) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 870.025822] env[62569]: DEBUG nova.network.neutron [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 870.067955] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.462s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.068536] env[62569]: INFO nova.compute.manager [None req-e91b4932-4402-4dd2-9ffd-16d38f45fcd2 tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Successfully reverted task state from rebuilding on failure for instance. [ 870.074522] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.194s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.076595] env[62569]: INFO nova.compute.claims [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 870.194191] env[62569]: DEBUG nova.policy [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ab6d20862c54432cbafdda33ccc974ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '530ca7e7924743ab91a362a064a3111b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 870.206308] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.329820] env[62569]: DEBUG nova.compute.manager [req-3766cdaf-81ad-4c98-9b1e-477dbed600f2 req-466b03a6-d4c7-4277-9692-6b4f16cb73c9 service nova] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Received event network-changed-f33064d8-b284-4ed3-9a2b-01f27371be83 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 870.331237] env[62569]: DEBUG nova.compute.manager [req-3766cdaf-81ad-4c98-9b1e-477dbed600f2 req-466b03a6-d4c7-4277-9692-6b4f16cb73c9 service nova] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Refreshing instance network info cache due to event network-changed-f33064d8-b284-4ed3-9a2b-01f27371be83. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 870.331237] env[62569]: DEBUG oslo_concurrency.lockutils [req-3766cdaf-81ad-4c98-9b1e-477dbed600f2 req-466b03a6-d4c7-4277-9692-6b4f16cb73c9 service nova] Acquiring lock "refresh_cache-298efba9-226d-4105-8b32-76b3dddddb0b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.331237] env[62569]: DEBUG oslo_concurrency.lockutils [req-3766cdaf-81ad-4c98-9b1e-477dbed600f2 req-466b03a6-d4c7-4277-9692-6b4f16cb73c9 service nova] Acquired lock "refresh_cache-298efba9-226d-4105-8b32-76b3dddddb0b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.331432] env[62569]: DEBUG nova.network.neutron [req-3766cdaf-81ad-4c98-9b1e-477dbed600f2 req-466b03a6-d4c7-4277-9692-6b4f16cb73c9 service nova] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Refreshing network info cache for port f33064d8-b284-4ed3-9a2b-01f27371be83 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 870.345269] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250247, 'name': CreateVM_Task, 'duration_secs': 0.359135} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.345269] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 870.345269] env[62569]: DEBUG oslo_concurrency.lockutils [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.345269] env[62569]: DEBUG oslo_concurrency.lockutils [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.345269] env[62569]: DEBUG oslo_concurrency.lockutils [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 870.345269] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e61752e0-6d92-40c4-b3f0-0d353b3e2803 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.351542] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Waiting for the task: (returnval){ [ 870.351542] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5252d4fc-e4e1-dbf8-eb28-54c13b6be849" [ 870.351542] env[62569]: _type = "Task" [ 870.351542] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.363228] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5252d4fc-e4e1-dbf8-eb28-54c13b6be849, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.771022] env[62569]: DEBUG oslo_concurrency.lockutils [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquiring lock "refresh_cache-337d88a2-30b9-4846-929e-042bd7a64a65" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.771504] env[62569]: DEBUG oslo_concurrency.lockutils [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquired lock "refresh_cache-337d88a2-30b9-4846-929e-042bd7a64a65" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.771504] env[62569]: DEBUG nova.network.neutron [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 870.869743] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5252d4fc-e4e1-dbf8-eb28-54c13b6be849, 'name': SearchDatastore_Task, 'duration_secs': 0.019797} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.869970] env[62569]: DEBUG oslo_concurrency.lockutils [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.870326] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 870.870881] env[62569]: DEBUG oslo_concurrency.lockutils [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.871258] env[62569]: DEBUG oslo_concurrency.lockutils [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.871753] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 870.872209] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-88e5917c-bac1-40eb-87e8-bc4bc9a20820 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.882894] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 870.883266] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 870.886916] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a29d52ff-691e-47b3-b53f-cc6b58f47007 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.893458] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Waiting for the task: (returnval){ [ 870.893458] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5240500b-14a4-d1f1-6412-00f9367155da" [ 870.893458] env[62569]: _type = "Task" [ 870.893458] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.901751] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5240500b-14a4-d1f1-6412-00f9367155da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.181097] env[62569]: DEBUG nova.network.neutron [req-3766cdaf-81ad-4c98-9b1e-477dbed600f2 req-466b03a6-d4c7-4277-9692-6b4f16cb73c9 service nova] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Updated VIF entry in instance network info cache for port f33064d8-b284-4ed3-9a2b-01f27371be83. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 871.181909] env[62569]: DEBUG nova.network.neutron [req-3766cdaf-81ad-4c98-9b1e-477dbed600f2 req-466b03a6-d4c7-4277-9692-6b4f16cb73c9 service nova] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Updating instance_info_cache with network_info: [{"id": "f33064d8-b284-4ed3-9a2b-01f27371be83", "address": "fa:16:3e:0b:05:c4", "network": {"id": "8a82d478-4aa8-4179-b974-51200b4e2bc6", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1314992403-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1af23d0979b84b85b348b3868379ef2b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf33064d8-b2", "ovs_interfaceid": "f33064d8-b284-4ed3-9a2b-01f27371be83", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.309890] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4380743e-c370-4f47-8e58-054bd06bc252 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.321892] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-363fff97-dcbe-46e5-95c2-6e64e6f57332 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.355591] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94bfb11b-1170-487f-810b-4e0bfc9b21f0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.367141] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6416ef27-bab6-4cb2-a793-d87ccb2f97b1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.386061] env[62569]: DEBUG nova.compute.provider_tree [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 871.408267] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5240500b-14a4-d1f1-6412-00f9367155da, 'name': SearchDatastore_Task, 'duration_secs': 0.01465} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.409559] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27bf212a-e2e7-43f2-a8da-3d4ac61a0646 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.415879] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Waiting for the task: (returnval){ [ 871.415879] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52756a17-3c24-b407-8cc2-d47120c45f67" [ 871.415879] env[62569]: _type = "Task" [ 871.415879] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.418468] env[62569]: DEBUG nova.compute.manager [req-bccef79c-23ae-4560-b3a6-c0b2607fe9a4 req-3338eea8-748e-4a80-8391-d18739860a49 service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Received event network-vif-plugged-303f46d4-6e0a-418c-b62b-17323ab645b6 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 871.418715] env[62569]: DEBUG oslo_concurrency.lockutils [req-bccef79c-23ae-4560-b3a6-c0b2607fe9a4 req-3338eea8-748e-4a80-8391-d18739860a49 service nova] Acquiring lock "337d88a2-30b9-4846-929e-042bd7a64a65-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.419505] env[62569]: DEBUG oslo_concurrency.lockutils [req-bccef79c-23ae-4560-b3a6-c0b2607fe9a4 req-3338eea8-748e-4a80-8391-d18739860a49 service nova] Lock "337d88a2-30b9-4846-929e-042bd7a64a65-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.419505] env[62569]: DEBUG oslo_concurrency.lockutils [req-bccef79c-23ae-4560-b3a6-c0b2607fe9a4 req-3338eea8-748e-4a80-8391-d18739860a49 service nova] Lock "337d88a2-30b9-4846-929e-042bd7a64a65-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.419505] env[62569]: DEBUG nova.compute.manager [req-bccef79c-23ae-4560-b3a6-c0b2607fe9a4 req-3338eea8-748e-4a80-8391-d18739860a49 service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] No waiting events found dispatching network-vif-plugged-303f46d4-6e0a-418c-b62b-17323ab645b6 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 871.419505] env[62569]: WARNING nova.compute.manager [req-bccef79c-23ae-4560-b3a6-c0b2607fe9a4 req-3338eea8-748e-4a80-8391-d18739860a49 service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Received unexpected event network-vif-plugged-303f46d4-6e0a-418c-b62b-17323ab645b6 for instance with vm_state shelved_offloaded and task_state spawning. [ 871.429288] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52756a17-3c24-b407-8cc2-d47120c45f67, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.551777] env[62569]: DEBUG nova.network.neutron [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Updating instance_info_cache with network_info: [{"id": "303f46d4-6e0a-418c-b62b-17323ab645b6", "address": "fa:16:3e:68:ca:b8", "network": {"id": "497ea2f3-6685-4d07-9952-b5ccf3c0b2f8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-561818925-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c69ac971dbd419f870a9fa68850cb1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7654928b-7afe-42e3-a18d-68ecc775cefe", "external-id": "cl2-zone-807", "segmentation_id": 807, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap303f46d4-6e", "ovs_interfaceid": "303f46d4-6e0a-418c-b62b-17323ab645b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.642500] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Acquiring lock "a6ee3547-3a3d-4480-b451-d1fddd829b39" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.642824] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Lock "a6ee3547-3a3d-4480-b451-d1fddd829b39" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.685567] env[62569]: DEBUG oslo_concurrency.lockutils [req-3766cdaf-81ad-4c98-9b1e-477dbed600f2 req-466b03a6-d4c7-4277-9692-6b4f16cb73c9 service nova] Releasing lock "refresh_cache-298efba9-226d-4105-8b32-76b3dddddb0b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.911849] env[62569]: ERROR nova.scheduler.client.report [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [req-e24658b3-fd1b-4d28-aad4-a71f927d9c77] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID fa06556a-5785-4014-b8bd-bc240a0cf716. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e24658b3-fd1b-4d28-aad4-a71f927d9c77"}]} [ 871.936228] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52756a17-3c24-b407-8cc2-d47120c45f67, 'name': SearchDatastore_Task, 'duration_secs': 0.016481} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.936598] env[62569]: DEBUG nova.scheduler.client.report [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Refreshing inventories for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 871.940329] env[62569]: DEBUG oslo_concurrency.lockutils [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.940329] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 298efba9-226d-4105-8b32-76b3dddddb0b/298efba9-226d-4105-8b32-76b3dddddb0b.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 871.940329] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7e43d8ba-b9f2-40a6-84b1-eed1f5071895 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.950490] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Waiting for the task: (returnval){ [ 871.950490] env[62569]: value = "task-1250249" [ 871.950490] env[62569]: _type = "Task" [ 871.950490] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.963956] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Task: {'id': task-1250249, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.965635] env[62569]: DEBUG nova.scheduler.client.report [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Updating ProviderTree inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 871.967303] env[62569]: DEBUG nova.compute.provider_tree [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 871.984250] env[62569]: DEBUG nova.scheduler.client.report [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Refreshing aggregate associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, aggregates: None {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 872.010298] env[62569]: DEBUG nova.scheduler.client.report [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Refreshing trait associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 872.054139] env[62569]: DEBUG oslo_concurrency.lockutils [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Releasing lock "refresh_cache-337d88a2-30b9-4846-929e-042bd7a64a65" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.089065] env[62569]: DEBUG nova.virt.hardware [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='00d4354d7298fd8d4ed13a2dbf64f9ea',container_format='bare',created_at=2024-10-10T15:37:23Z,direct_url=,disk_format='vmdk',id=45f3d90c-b69b-4905-993e-439633bfeb29,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-883950706-shelved',owner='8c69ac971dbd419f870a9fa68850cb1c',properties=ImageMetaProps,protected=,size=31670272,status='active',tags=,updated_at=2024-10-10T15:37:37Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 872.089408] env[62569]: DEBUG nova.virt.hardware [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 872.089496] env[62569]: DEBUG nova.virt.hardware [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 872.089681] env[62569]: DEBUG nova.virt.hardware [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 872.089853] env[62569]: DEBUG nova.virt.hardware [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 872.090092] env[62569]: DEBUG nova.virt.hardware [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 872.090501] env[62569]: DEBUG nova.virt.hardware [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 872.090705] env[62569]: DEBUG nova.virt.hardware [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 872.090920] env[62569]: DEBUG nova.virt.hardware [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 872.091142] env[62569]: DEBUG nova.virt.hardware [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 872.091332] env[62569]: DEBUG nova.virt.hardware [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 872.092520] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad0f9855-a0a1-4d50-8207-67b6b1afdcbb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.096312] env[62569]: DEBUG nova.network.neutron [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Successfully updated port: 62c1052e-dc0a-4fb6-a399-3e238bf93789 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 872.110290] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-281283a3-051e-44c7-bd1d-f3009e2eee28 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.130840] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:68:ca:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7654928b-7afe-42e3-a18d-68ecc775cefe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '303f46d4-6e0a-418c-b62b-17323ab645b6', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 872.139497] env[62569]: DEBUG oslo.service.loopingcall [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.143399] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 872.144064] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d9c1161f-66dd-4199-965c-015dc6c5b033 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.166505] env[62569]: DEBUG nova.compute.manager [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 872.175808] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 872.175808] env[62569]: value = "task-1250250" [ 872.175808] env[62569]: _type = "Task" [ 872.175808] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.186902] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250250, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.292349] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ddce221-47f8-4dc5-b737-a688b4d87986 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.301967] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b5f7ab-9554-4d68-83e6-aede57acf602 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.338803] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-078afc22-157d-4d3c-96b9-8603aaceef7b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.346789] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa6aa8f6-6eb7-4343-81da-fbbc5fefbb9c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.364762] env[62569]: DEBUG nova.compute.provider_tree [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 872.460210] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Task: {'id': task-1250249, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.513484] env[62569]: DEBUG nova.compute.manager [req-530f4d51-4cec-41eb-9f37-01d4882772d8 req-9d709a15-905a-4809-b26e-c4b614a90bc0 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Received event network-vif-plugged-62c1052e-dc0a-4fb6-a399-3e238bf93789 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 872.513737] env[62569]: DEBUG oslo_concurrency.lockutils [req-530f4d51-4cec-41eb-9f37-01d4882772d8 req-9d709a15-905a-4809-b26e-c4b614a90bc0 service nova] Acquiring lock "807d3025-d6a7-4778-a829-a61e2c7495c9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.513808] env[62569]: DEBUG oslo_concurrency.lockutils [req-530f4d51-4cec-41eb-9f37-01d4882772d8 req-9d709a15-905a-4809-b26e-c4b614a90bc0 service nova] Lock "807d3025-d6a7-4778-a829-a61e2c7495c9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.517227] env[62569]: DEBUG oslo_concurrency.lockutils [req-530f4d51-4cec-41eb-9f37-01d4882772d8 req-9d709a15-905a-4809-b26e-c4b614a90bc0 service nova] Lock "807d3025-d6a7-4778-a829-a61e2c7495c9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.517227] env[62569]: DEBUG nova.compute.manager [req-530f4d51-4cec-41eb-9f37-01d4882772d8 req-9d709a15-905a-4809-b26e-c4b614a90bc0 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] No waiting events found dispatching network-vif-plugged-62c1052e-dc0a-4fb6-a399-3e238bf93789 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 872.517227] env[62569]: WARNING nova.compute.manager [req-530f4d51-4cec-41eb-9f37-01d4882772d8 req-9d709a15-905a-4809-b26e-c4b614a90bc0 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Received unexpected event network-vif-plugged-62c1052e-dc0a-4fb6-a399-3e238bf93789 for instance with vm_state active and task_state None. [ 872.517227] env[62569]: DEBUG nova.compute.manager [req-530f4d51-4cec-41eb-9f37-01d4882772d8 req-9d709a15-905a-4809-b26e-c4b614a90bc0 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Received event network-changed-62c1052e-dc0a-4fb6-a399-3e238bf93789 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 872.517227] env[62569]: DEBUG nova.compute.manager [req-530f4d51-4cec-41eb-9f37-01d4882772d8 req-9d709a15-905a-4809-b26e-c4b614a90bc0 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Refreshing instance network info cache due to event network-changed-62c1052e-dc0a-4fb6-a399-3e238bf93789. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 872.517227] env[62569]: DEBUG oslo_concurrency.lockutils [req-530f4d51-4cec-41eb-9f37-01d4882772d8 req-9d709a15-905a-4809-b26e-c4b614a90bc0 service nova] Acquiring lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.517227] env[62569]: DEBUG oslo_concurrency.lockutils [req-530f4d51-4cec-41eb-9f37-01d4882772d8 req-9d709a15-905a-4809-b26e-c4b614a90bc0 service nova] Acquired lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.517227] env[62569]: DEBUG nova.network.neutron [req-530f4d51-4cec-41eb-9f37-01d4882772d8 req-9d709a15-905a-4809-b26e-c4b614a90bc0 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Refreshing network info cache for port 62c1052e-dc0a-4fb6-a399-3e238bf93789 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 872.604035] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.666623] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1f3a574-b618-4a13-86f3-686b2e7db7f7 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Volume attach. Driver type: vmdk {{(pid=62569) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 872.666893] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1f3a574-b618-4a13-86f3-686b2e7db7f7 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269446', 'volume_id': 'f29f7945-6df2-4fc4-bd59-5d5daee539e7', 'name': 'volume-f29f7945-6df2-4fc4-bd59-5d5daee539e7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cb9d8991-b7ba-4bf7-9d22-1391f4cea768', 'attached_at': '', 'detached_at': '', 'volume_id': 'f29f7945-6df2-4fc4-bd59-5d5daee539e7', 'serial': 'f29f7945-6df2-4fc4-bd59-5d5daee539e7'} {{(pid=62569) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 872.667971] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ec0fb4-af8f-4f84-9e71-ca0aaa2a3258 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.699245] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdb85439-49a0-4aa6-8832-17791e5b2ee6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.725375] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250250, 'name': CreateVM_Task, 'duration_secs': 0.38089} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.733017] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1f3a574-b618-4a13-86f3-686b2e7db7f7 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] volume-f29f7945-6df2-4fc4-bd59-5d5daee539e7/volume-f29f7945-6df2-4fc4-bd59-5d5daee539e7.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 872.735062] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.735378] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 872.735692] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c33f3dd8-561a-43dc-9200-109478f8dae8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.755181] env[62569]: DEBUG oslo_concurrency.lockutils [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/45f3d90c-b69b-4905-993e-439633bfeb29" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.755181] env[62569]: DEBUG oslo_concurrency.lockutils [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquired lock "[datastore2] devstack-image-cache_base/45f3d90c-b69b-4905-993e-439633bfeb29" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.755181] env[62569]: DEBUG oslo_concurrency.lockutils [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/45f3d90c-b69b-4905-993e-439633bfeb29" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 872.755181] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb8e0fc3-7bf8-47d5-9e57-19f2f4974ff6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.759727] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 872.759727] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5247a074-5d86-8002-1ba9-7ab6b0cda454" [ 872.759727] env[62569]: _type = "Task" [ 872.759727] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.765515] env[62569]: DEBUG oslo_vmware.api [None req-f1f3a574-b618-4a13-86f3-686b2e7db7f7 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 872.765515] env[62569]: value = "task-1250251" [ 872.765515] env[62569]: _type = "Task" [ 872.765515] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.772827] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5247a074-5d86-8002-1ba9-7ab6b0cda454, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.778375] env[62569]: DEBUG oslo_vmware.api [None req-f1f3a574-b618-4a13-86f3-686b2e7db7f7 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250251, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.902965] env[62569]: DEBUG nova.scheduler.client.report [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Updated inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with generation 99 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 872.903272] env[62569]: DEBUG nova.compute.provider_tree [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Updating resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 generation from 99 to 100 during operation: update_inventory {{(pid=62569) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 872.903451] env[62569]: DEBUG nova.compute.provider_tree [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 872.962626] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Task: {'id': task-1250249, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.530095} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.963050] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 298efba9-226d-4105-8b32-76b3dddddb0b/298efba9-226d-4105-8b32-76b3dddddb0b.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 872.963268] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 872.963586] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d7d28501-8aab-48fe-99ce-1b0b067fb8c9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.972100] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Waiting for the task: (returnval){ [ 872.972100] env[62569]: value = "task-1250252" [ 872.972100] env[62569]: _type = "Task" [ 872.972100] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.985047] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Task: {'id': task-1250252, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.276046] env[62569]: DEBUG nova.network.neutron [req-530f4d51-4cec-41eb-9f37-01d4882772d8 req-9d709a15-905a-4809-b26e-c4b614a90bc0 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Added VIF to instance network info cache for port 62c1052e-dc0a-4fb6-a399-3e238bf93789. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3546}} [ 873.276537] env[62569]: DEBUG nova.network.neutron [req-530f4d51-4cec-41eb-9f37-01d4882772d8 req-9d709a15-905a-4809-b26e-c4b614a90bc0 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Updating instance_info_cache with network_info: [{"id": "9d855102-33ab-4574-b791-dfa6fb5b0f09", "address": "fa:16:3e:b3:4b:f5", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d855102-33", "ovs_interfaceid": "9d855102-33ab-4574-b791-dfa6fb5b0f09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18", "address": "fa:16:3e:df:90:ac", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f2ad2ec-3e", "ovs_interfaceid": "5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "62c1052e-dc0a-4fb6-a399-3e238bf93789", "address": "fa:16:3e:9d:69:9f", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62c1052e-dc", "ovs_interfaceid": "62c1052e-dc0a-4fb6-a399-3e238bf93789", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.278460] env[62569]: DEBUG oslo_concurrency.lockutils [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Releasing lock "[datastore2] devstack-image-cache_base/45f3d90c-b69b-4905-993e-439633bfeb29" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.278714] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Processing image 45f3d90c-b69b-4905-993e-439633bfeb29 {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 873.279644] env[62569]: DEBUG oslo_concurrency.lockutils [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/45f3d90c-b69b-4905-993e-439633bfeb29/45f3d90c-b69b-4905-993e-439633bfeb29.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.279644] env[62569]: DEBUG oslo_concurrency.lockutils [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquired lock "[datastore2] devstack-image-cache_base/45f3d90c-b69b-4905-993e-439633bfeb29/45f3d90c-b69b-4905-993e-439633bfeb29.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.279644] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 873.288018] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-92f771e5-b901-45a4-aea4-5fa6b64d9f33 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.293250] env[62569]: DEBUG oslo_vmware.api [None req-f1f3a574-b618-4a13-86f3-686b2e7db7f7 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250251, 'name': ReconfigVM_Task, 'duration_secs': 0.499636} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.293565] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1f3a574-b618-4a13-86f3-686b2e7db7f7 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Reconfigured VM instance instance-00000048 to attach disk [datastore2] volume-f29f7945-6df2-4fc4-bd59-5d5daee539e7/volume-f29f7945-6df2-4fc4-bd59-5d5daee539e7.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 873.303241] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-efce660a-2bf0-4ab5-b5c3-13226e70d1c2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.317478] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 873.317717] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 873.319648] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f147fe2d-3f6b-451f-ac4f-389c36204bb6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.325531] env[62569]: DEBUG oslo_vmware.api [None req-f1f3a574-b618-4a13-86f3-686b2e7db7f7 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 873.325531] env[62569]: value = "task-1250253" [ 873.325531] env[62569]: _type = "Task" [ 873.325531] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.329353] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 873.329353] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ba6849-2422-2cae-4b61-8b80d561dc3a" [ 873.329353] env[62569]: _type = "Task" [ 873.329353] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.337233] env[62569]: DEBUG oslo_vmware.api [None req-f1f3a574-b618-4a13-86f3-686b2e7db7f7 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250253, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.347341] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ba6849-2422-2cae-4b61-8b80d561dc3a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.410516] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.335s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.411265] env[62569]: DEBUG nova.compute.manager [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 873.414561] env[62569]: DEBUG oslo_concurrency.lockutils [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.813s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.414626] env[62569]: DEBUG nova.objects.instance [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Lazy-loading 'resources' on Instance uuid fdeac8b8-a34d-4b4c-8f72-281f84e22ebe {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 873.482179] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Task: {'id': task-1250252, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074729} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.482470] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 873.483317] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20c7f314-3d6b-4842-ad0d-dcfe956c9ace {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.498753] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4322fc1b-e948-40a0-afe0-a798015b8c51 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.499252] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4322fc1b-e948-40a0-afe0-a798015b8c51 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.499424] env[62569]: DEBUG nova.compute.manager [None req-4322fc1b-e948-40a0-afe0-a798015b8c51 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 873.509924] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] 298efba9-226d-4105-8b32-76b3dddddb0b/298efba9-226d-4105-8b32-76b3dddddb0b.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 873.511258] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e177ed7a-ec62-4271-8909-72feece8b266 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.514977] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-019d16f4-7b39-4320-a65c-ccd02b452789 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.536217] env[62569]: DEBUG nova.compute.manager [None req-4322fc1b-e948-40a0-afe0-a798015b8c51 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62569) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 873.536846] env[62569]: DEBUG nova.objects.instance [None req-4322fc1b-e948-40a0-afe0-a798015b8c51 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lazy-loading 'flavor' on Instance uuid e4eadcdf-a04a-4255-ba1c-fe20156c655f {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 873.539322] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Waiting for the task: (returnval){ [ 873.539322] env[62569]: value = "task-1250254" [ 873.539322] env[62569]: _type = "Task" [ 873.539322] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.551696] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Task: {'id': task-1250254, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.781923] env[62569]: DEBUG oslo_concurrency.lockutils [req-530f4d51-4cec-41eb-9f37-01d4882772d8 req-9d709a15-905a-4809-b26e-c4b614a90bc0 service nova] Releasing lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.781923] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.782483] env[62569]: DEBUG nova.network.neutron [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 873.839855] env[62569]: DEBUG oslo_vmware.api [None req-f1f3a574-b618-4a13-86f3-686b2e7db7f7 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250253, 'name': ReconfigVM_Task, 'duration_secs': 0.184676} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.845184] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1f3a574-b618-4a13-86f3-686b2e7db7f7 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269446', 'volume_id': 'f29f7945-6df2-4fc4-bd59-5d5daee539e7', 'name': 'volume-f29f7945-6df2-4fc4-bd59-5d5daee539e7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cb9d8991-b7ba-4bf7-9d22-1391f4cea768', 'attached_at': '', 'detached_at': '', 'volume_id': 'f29f7945-6df2-4fc4-bd59-5d5daee539e7', 'serial': 'f29f7945-6df2-4fc4-bd59-5d5daee539e7'} {{(pid=62569) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 873.848799] env[62569]: DEBUG nova.compute.manager [req-d24ee957-f65e-481a-9cdb-068f0e5cada5 req-f6deb82a-7bb4-4b1d-b8c1-70b4fb9532b3 service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Received event network-changed-303f46d4-6e0a-418c-b62b-17323ab645b6 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 873.848982] env[62569]: DEBUG nova.compute.manager [req-d24ee957-f65e-481a-9cdb-068f0e5cada5 req-f6deb82a-7bb4-4b1d-b8c1-70b4fb9532b3 service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Refreshing instance network info cache due to event network-changed-303f46d4-6e0a-418c-b62b-17323ab645b6. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 873.849216] env[62569]: DEBUG oslo_concurrency.lockutils [req-d24ee957-f65e-481a-9cdb-068f0e5cada5 req-f6deb82a-7bb4-4b1d-b8c1-70b4fb9532b3 service nova] Acquiring lock "refresh_cache-337d88a2-30b9-4846-929e-042bd7a64a65" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.849409] env[62569]: DEBUG oslo_concurrency.lockutils [req-d24ee957-f65e-481a-9cdb-068f0e5cada5 req-f6deb82a-7bb4-4b1d-b8c1-70b4fb9532b3 service nova] Acquired lock "refresh_cache-337d88a2-30b9-4846-929e-042bd7a64a65" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.849551] env[62569]: DEBUG nova.network.neutron [req-d24ee957-f65e-481a-9cdb-068f0e5cada5 req-f6deb82a-7bb4-4b1d-b8c1-70b4fb9532b3 service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Refreshing network info cache for port 303f46d4-6e0a-418c-b62b-17323ab645b6 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 873.851457] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Preparing fetch location {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 873.851853] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Fetch image to [datastore2] OSTACK_IMG_ee1abb3b-b3e0-49fa-8bb4-04290a98ab3c/OSTACK_IMG_ee1abb3b-b3e0-49fa-8bb4-04290a98ab3c.vmdk {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 873.852030] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Downloading stream optimized image 45f3d90c-b69b-4905-993e-439633bfeb29 to [datastore2] OSTACK_IMG_ee1abb3b-b3e0-49fa-8bb4-04290a98ab3c/OSTACK_IMG_ee1abb3b-b3e0-49fa-8bb4-04290a98ab3c.vmdk on the data store datastore2 as vApp {{(pid=62569) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 873.852453] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Downloading image file data 45f3d90c-b69b-4905-993e-439633bfeb29 to the ESX as VM named 'OSTACK_IMG_ee1abb3b-b3e0-49fa-8bb4-04290a98ab3c' {{(pid=62569) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 873.918822] env[62569]: DEBUG nova.compute.utils [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 873.924081] env[62569]: DEBUG nova.compute.manager [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 873.924428] env[62569]: DEBUG nova.network.neutron [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 873.927790] env[62569]: DEBUG oslo_vmware.rw_handles [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 873.927790] env[62569]: value = "resgroup-9" [ 873.927790] env[62569]: _type = "ResourcePool" [ 873.927790] env[62569]: }. {{(pid=62569) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 873.928127] env[62569]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-4bad8726-b6ff-4e80-bab9-d13bfa055c62 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.956688] env[62569]: DEBUG oslo_vmware.rw_handles [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lease: (returnval){ [ 873.956688] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52c329e7-8906-4893-61d6-d517aa2f62fd" [ 873.956688] env[62569]: _type = "HttpNfcLease" [ 873.956688] env[62569]: } obtained for vApp import into resource pool (val){ [ 873.956688] env[62569]: value = "resgroup-9" [ 873.956688] env[62569]: _type = "ResourcePool" [ 873.956688] env[62569]: }. {{(pid=62569) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 873.957071] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the lease: (returnval){ [ 873.957071] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52c329e7-8906-4893-61d6-d517aa2f62fd" [ 873.957071] env[62569]: _type = "HttpNfcLease" [ 873.957071] env[62569]: } to be ready. {{(pid=62569) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 873.965667] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 873.965667] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52c329e7-8906-4893-61d6-d517aa2f62fd" [ 873.965667] env[62569]: _type = "HttpNfcLease" [ 873.965667] env[62569]: } is initializing. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 873.995402] env[62569]: DEBUG nova.policy [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5da4f6b7a6784a73bd3fed04b275041b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bdba6022e3d4697a336ca28ca4eccec', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 874.055228] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Task: {'id': task-1250254, 'name': ReconfigVM_Task, 'duration_secs': 0.34558} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.055682] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Reconfigured VM instance instance-00000051 to attach disk [datastore2] 298efba9-226d-4105-8b32-76b3dddddb0b/298efba9-226d-4105-8b32-76b3dddddb0b.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 874.061699] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3d023f49-6154-4332-81c6-d5c2d6bb7e95 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.069718] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Waiting for the task: (returnval){ [ 874.069718] env[62569]: value = "task-1250256" [ 874.069718] env[62569]: _type = "Task" [ 874.069718] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.079105] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Task: {'id': task-1250256, 'name': Rename_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.189817] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24244783-8d88-4317-a5a0-5fde259973b9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.199109] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e907fdc5-0362-4f70-9ba5-a0628a02e434 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.236830] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20dcebae-5b63-468e-bb3c-b9411e1e8643 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.245175] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bdac075-bb85-4940-af89-cabc3e1a4f82 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.259871] env[62569]: DEBUG nova.compute.provider_tree [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 874.326703] env[62569]: WARNING nova.network.neutron [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] 29fcbfe8-35ce-4e8d-874b-2b14ba720adf already exists in list: networks containing: ['29fcbfe8-35ce-4e8d-874b-2b14ba720adf']. ignoring it [ 874.326930] env[62569]: WARNING nova.network.neutron [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] 29fcbfe8-35ce-4e8d-874b-2b14ba720adf already exists in list: networks containing: ['29fcbfe8-35ce-4e8d-874b-2b14ba720adf']. ignoring it [ 874.327125] env[62569]: WARNING nova.network.neutron [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] 29fcbfe8-35ce-4e8d-874b-2b14ba720adf already exists in list: networks containing: ['29fcbfe8-35ce-4e8d-874b-2b14ba720adf']. ignoring it [ 874.327365] env[62569]: WARNING nova.network.neutron [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] 62c1052e-dc0a-4fb6-a399-3e238bf93789 already exists in list: port_ids containing: ['62c1052e-dc0a-4fb6-a399-3e238bf93789']. ignoring it [ 874.427100] env[62569]: DEBUG nova.compute.manager [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 874.450535] env[62569]: DEBUG nova.network.neutron [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Successfully created port: 19f8c74e-ebd3-4b14-9e7b-1539ce706834 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 874.467893] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 874.467893] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52c329e7-8906-4893-61d6-d517aa2f62fd" [ 874.467893] env[62569]: _type = "HttpNfcLease" [ 874.467893] env[62569]: } is initializing. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 874.557423] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4322fc1b-e948-40a0-afe0-a798015b8c51 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 874.557908] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8ce22e41-a703-4c5c-9118-fc08e91b181e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.566194] env[62569]: DEBUG oslo_vmware.api [None req-4322fc1b-e948-40a0-afe0-a798015b8c51 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 874.566194] env[62569]: value = "task-1250257" [ 874.566194] env[62569]: _type = "Task" [ 874.566194] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.583187] env[62569]: DEBUG oslo_vmware.api [None req-4322fc1b-e948-40a0-afe0-a798015b8c51 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250257, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.586762] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Task: {'id': task-1250256, 'name': Rename_Task, 'duration_secs': 0.174177} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.589279] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 874.589563] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fe361cb3-2d68-4c3b-a6d7-74fc3eb76cd5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.597355] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Waiting for the task: (returnval){ [ 874.597355] env[62569]: value = "task-1250258" [ 874.597355] env[62569]: _type = "Task" [ 874.597355] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.609164] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Task: {'id': task-1250258, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.799131] env[62569]: DEBUG nova.network.neutron [req-d24ee957-f65e-481a-9cdb-068f0e5cada5 req-f6deb82a-7bb4-4b1d-b8c1-70b4fb9532b3 service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Updated VIF entry in instance network info cache for port 303f46d4-6e0a-418c-b62b-17323ab645b6. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 874.799131] env[62569]: DEBUG nova.network.neutron [req-d24ee957-f65e-481a-9cdb-068f0e5cada5 req-f6deb82a-7bb4-4b1d-b8c1-70b4fb9532b3 service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Updating instance_info_cache with network_info: [{"id": "303f46d4-6e0a-418c-b62b-17323ab645b6", "address": "fa:16:3e:68:ca:b8", "network": {"id": "497ea2f3-6685-4d07-9952-b5ccf3c0b2f8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-561818925-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c69ac971dbd419f870a9fa68850cb1c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7654928b-7afe-42e3-a18d-68ecc775cefe", "external-id": "cl2-zone-807", "segmentation_id": 807, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap303f46d4-6e", "ovs_interfaceid": "303f46d4-6e0a-418c-b62b-17323ab645b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.809170] env[62569]: DEBUG nova.scheduler.client.report [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Updated inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with generation 100 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 874.809493] env[62569]: DEBUG nova.compute.provider_tree [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Updating resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 generation from 100 to 101 during operation: update_inventory {{(pid=62569) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 874.809694] env[62569]: DEBUG nova.compute.provider_tree [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 874.897364] env[62569]: DEBUG nova.objects.instance [None req-f1f3a574-b618-4a13-86f3-686b2e7db7f7 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lazy-loading 'flavor' on Instance uuid cb9d8991-b7ba-4bf7-9d22-1391f4cea768 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.967708] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 874.967708] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52c329e7-8906-4893-61d6-d517aa2f62fd" [ 874.967708] env[62569]: _type = "HttpNfcLease" [ 874.967708] env[62569]: } is ready. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 874.968202] env[62569]: DEBUG oslo_vmware.rw_handles [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 874.968202] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52c329e7-8906-4893-61d6-d517aa2f62fd" [ 874.968202] env[62569]: _type = "HttpNfcLease" [ 874.968202] env[62569]: }. {{(pid=62569) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 874.968961] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70555a9f-f829-41a9-a778-22dcca213c13 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.979930] env[62569]: DEBUG oslo_vmware.rw_handles [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526f5442-0ad9-543f-3f57-8f0025cd71e6/disk-0.vmdk from lease info. {{(pid=62569) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 874.980208] env[62569]: DEBUG oslo_vmware.rw_handles [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Creating HTTP connection to write to file with size = 31670272 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526f5442-0ad9-543f-3f57-8f0025cd71e6/disk-0.vmdk. {{(pid=62569) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 875.044720] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-dc113bd7-6af6-4242-9924-89a25f1442d7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.077716] env[62569]: DEBUG oslo_vmware.api [None req-4322fc1b-e948-40a0-afe0-a798015b8c51 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250257, 'name': PowerOffVM_Task, 'duration_secs': 0.203874} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.078016] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4322fc1b-e948-40a0-afe0-a798015b8c51 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 875.078229] env[62569]: DEBUG nova.compute.manager [None req-4322fc1b-e948-40a0-afe0-a798015b8c51 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 875.079103] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aae5d332-0e9d-451e-abce-88efe36abc3e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.107297] env[62569]: DEBUG oslo_vmware.api [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Task: {'id': task-1250258, 'name': PowerOnVM_Task, 'duration_secs': 0.503884} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.107297] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 875.107479] env[62569]: INFO nova.compute.manager [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Took 8.69 seconds to spawn the instance on the hypervisor. [ 875.108264] env[62569]: DEBUG nova.compute.manager [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 875.108518] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dab8b6d-f411-4fdc-ba96-96ced98b2c7d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.199520] env[62569]: DEBUG nova.network.neutron [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Updating instance_info_cache with network_info: [{"id": "9d855102-33ab-4574-b791-dfa6fb5b0f09", "address": "fa:16:3e:b3:4b:f5", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d855102-33", "ovs_interfaceid": "9d855102-33ab-4574-b791-dfa6fb5b0f09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18", "address": "fa:16:3e:df:90:ac", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f2ad2ec-3e", "ovs_interfaceid": "5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "62c1052e-dc0a-4fb6-a399-3e238bf93789", "address": "fa:16:3e:9d:69:9f", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62c1052e-dc", "ovs_interfaceid": "62c1052e-dc0a-4fb6-a399-3e238bf93789", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.302144] env[62569]: DEBUG oslo_concurrency.lockutils [req-d24ee957-f65e-481a-9cdb-068f0e5cada5 req-f6deb82a-7bb4-4b1d-b8c1-70b4fb9532b3 service nova] Releasing lock "refresh_cache-337d88a2-30b9-4846-929e-042bd7a64a65" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.316547] env[62569]: DEBUG oslo_concurrency.lockutils [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.902s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.321168] env[62569]: DEBUG oslo_concurrency.lockutils [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 6.481s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.404426] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f1f3a574-b618-4a13-86f3-686b2e7db7f7 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cb9d8991-b7ba-4bf7-9d22-1391f4cea768" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.422s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.441556] env[62569]: DEBUG nova.compute.manager [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 875.593619] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4322fc1b-e948-40a0-afe0-a798015b8c51 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.094s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.629104] env[62569]: INFO nova.compute.manager [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Took 17.46 seconds to build instance. [ 875.702878] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.703638] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.704010] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.704909] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e30551a-17f4-4849-aa5f-3828f20f8d60 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.731838] env[62569]: DEBUG nova.virt.hardware [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 875.732142] env[62569]: DEBUG nova.virt.hardware [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 875.732313] env[62569]: DEBUG nova.virt.hardware [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 875.732509] env[62569]: DEBUG nova.virt.hardware [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 875.732676] env[62569]: DEBUG nova.virt.hardware [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 875.732845] env[62569]: DEBUG nova.virt.hardware [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 875.733078] env[62569]: DEBUG nova.virt.hardware [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 875.733254] env[62569]: DEBUG nova.virt.hardware [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 875.733430] env[62569]: DEBUG nova.virt.hardware [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 875.734112] env[62569]: DEBUG nova.virt.hardware [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 875.734371] env[62569]: DEBUG nova.virt.hardware [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 875.741242] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Reconfiguring VM to attach interface {{(pid=62569) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 875.745934] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-240bf0c2-73fa-43c5-9652-2ef101b6fbbc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.767579] env[62569]: DEBUG oslo_vmware.api [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 875.767579] env[62569]: value = "task-1250259" [ 875.767579] env[62569]: _type = "Task" [ 875.767579] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.770550] env[62569]: DEBUG nova.virt.hardware [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 875.770987] env[62569]: DEBUG nova.virt.hardware [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 875.771214] env[62569]: DEBUG nova.virt.hardware [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 875.771441] env[62569]: DEBUG nova.virt.hardware [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 875.771670] env[62569]: DEBUG nova.virt.hardware [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 875.771902] env[62569]: DEBUG nova.virt.hardware [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 875.772102] env[62569]: DEBUG nova.virt.hardware [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 875.772288] env[62569]: DEBUG nova.virt.hardware [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 875.772517] env[62569]: DEBUG nova.virt.hardware [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 875.772716] env[62569]: DEBUG nova.virt.hardware [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 875.775166] env[62569]: DEBUG nova.virt.hardware [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 875.776187] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f3623b8-0513-461f-91cc-f406db62ad65 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.799051] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38259b6d-82a6-4207-bbfb-bd75c4756c8b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.803859] env[62569]: DEBUG oslo_vmware.api [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250259, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.829824] env[62569]: INFO nova.compute.claims [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 875.838143] env[62569]: DEBUG oslo_concurrency.lockutils [None req-57c96859-1fbf-46c2-8652-064fee27acee tempest-ServerActionsV293TestJSON-1177409275 tempest-ServerActionsV293TestJSON-1177409275-project-member] Lock "fdeac8b8-a34d-4b4c-8f72-281f84e22ebe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.495s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.133022] env[62569]: DEBUG oslo_concurrency.lockutils [None req-84f62364-03c2-4d50-8470-abd40899f8ac tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Lock "298efba9-226d-4105-8b32-76b3dddddb0b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.976s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.200990] env[62569]: DEBUG nova.network.neutron [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Successfully updated port: 19f8c74e-ebd3-4b14-9e7b-1539ce706834 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 876.286361] env[62569]: DEBUG oslo_vmware.api [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250259, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.336809] env[62569]: DEBUG oslo_vmware.rw_handles [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Completed reading data from the image iterator. {{(pid=62569) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 876.337147] env[62569]: DEBUG oslo_vmware.rw_handles [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526f5442-0ad9-543f-3f57-8f0025cd71e6/disk-0.vmdk. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 876.338287] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8106fe8-a8ae-415e-9f35-8b6763b58a5d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.343434] env[62569]: INFO nova.compute.resource_tracker [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Updating resource usage from migration 6322afb2-1484-48c6-8019-ecd2841d0d0b [ 876.351905] env[62569]: DEBUG oslo_vmware.rw_handles [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526f5442-0ad9-543f-3f57-8f0025cd71e6/disk-0.vmdk is in state: ready. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 876.353287] env[62569]: DEBUG oslo_vmware.rw_handles [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526f5442-0ad9-543f-3f57-8f0025cd71e6/disk-0.vmdk. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 876.353567] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-dfb23d13-77f2-44c2-a6e0-38caf10553e7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.560089] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82dd56e1-c1a5-4917-bff1-0d82dd138ca9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.568216] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f30689c0-d122-4459-aa0e-a92bb677f932 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.599011] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e2d249e-b698-4e23-9bb5-1f753befcbd1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.606779] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b41fff-a273-4b06-bd9b-1b0b67e2fe1c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.620480] env[62569]: DEBUG nova.compute.provider_tree [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.641415] env[62569]: DEBUG oslo_vmware.rw_handles [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526f5442-0ad9-543f-3f57-8f0025cd71e6/disk-0.vmdk. {{(pid=62569) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 876.641709] env[62569]: INFO nova.virt.vmwareapi.images [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Downloaded image file data 45f3d90c-b69b-4905-993e-439633bfeb29 [ 876.642595] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c26262b-1cd8-4674-b1f9-96bedba54e3f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.659734] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4945ce46-0d34-47d5-bf11-fa16baed9426 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.704155] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "refresh_cache-108d5bde-e463-4a69-816b-bdd3c03e13cd" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.704155] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired lock "refresh_cache-108d5bde-e463-4a69-816b-bdd3c03e13cd" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.704155] env[62569]: DEBUG nova.network.neutron [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 876.749175] env[62569]: INFO nova.virt.vmwareapi.images [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] The imported VM was unregistered [ 876.751145] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Caching image {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 876.751422] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Creating directory with path [datastore2] devstack-image-cache_base/45f3d90c-b69b-4905-993e-439633bfeb29 {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 876.752571] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-989420ef-1d31-4e43-844d-c9a8a5ec658e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.766143] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Created directory with path [datastore2] devstack-image-cache_base/45f3d90c-b69b-4905-993e-439633bfeb29 {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 876.766143] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_ee1abb3b-b3e0-49fa-8bb4-04290a98ab3c/OSTACK_IMG_ee1abb3b-b3e0-49fa-8bb4-04290a98ab3c.vmdk to [datastore2] devstack-image-cache_base/45f3d90c-b69b-4905-993e-439633bfeb29/45f3d90c-b69b-4905-993e-439633bfeb29.vmdk. {{(pid=62569) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 876.766345] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-3bcb5c9b-dd8e-456a-8b5e-bc67d6388063 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.773875] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 876.773875] env[62569]: value = "task-1250261" [ 876.773875] env[62569]: _type = "Task" [ 876.773875] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.786562] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250261, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.789790] env[62569]: DEBUG oslo_vmware.api [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250259, 'name': ReconfigVM_Task, 'duration_secs': 0.780544} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.790322] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.790639] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Reconfigured VM to attach interface {{(pid=62569) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 876.838162] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "534c9ece-c8df-4528-83d8-c6c0ef8c4793" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.838426] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "534c9ece-c8df-4528-83d8-c6c0ef8c4793" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.124244] env[62569]: DEBUG nova.scheduler.client.report [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 877.172885] env[62569]: DEBUG nova.compute.manager [req-eec4f092-2554-4460-a41f-b14284c91d74 req-69c309d9-bcdb-4ce1-acdd-117a50dbed0f service nova] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Received event network-vif-plugged-19f8c74e-ebd3-4b14-9e7b-1539ce706834 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 877.173156] env[62569]: DEBUG oslo_concurrency.lockutils [req-eec4f092-2554-4460-a41f-b14284c91d74 req-69c309d9-bcdb-4ce1-acdd-117a50dbed0f service nova] Acquiring lock "108d5bde-e463-4a69-816b-bdd3c03e13cd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.173371] env[62569]: DEBUG oslo_concurrency.lockutils [req-eec4f092-2554-4460-a41f-b14284c91d74 req-69c309d9-bcdb-4ce1-acdd-117a50dbed0f service nova] Lock "108d5bde-e463-4a69-816b-bdd3c03e13cd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.173540] env[62569]: DEBUG oslo_concurrency.lockutils [req-eec4f092-2554-4460-a41f-b14284c91d74 req-69c309d9-bcdb-4ce1-acdd-117a50dbed0f service nova] Lock "108d5bde-e463-4a69-816b-bdd3c03e13cd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.173794] env[62569]: DEBUG nova.compute.manager [req-eec4f092-2554-4460-a41f-b14284c91d74 req-69c309d9-bcdb-4ce1-acdd-117a50dbed0f service nova] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] No waiting events found dispatching network-vif-plugged-19f8c74e-ebd3-4b14-9e7b-1539ce706834 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 877.173983] env[62569]: WARNING nova.compute.manager [req-eec4f092-2554-4460-a41f-b14284c91d74 req-69c309d9-bcdb-4ce1-acdd-117a50dbed0f service nova] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Received unexpected event network-vif-plugged-19f8c74e-ebd3-4b14-9e7b-1539ce706834 for instance with vm_state building and task_state spawning. [ 877.238603] env[62569]: DEBUG nova.network.neutron [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 877.290942] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250261, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.296559] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bd2949c0-9c2b-4bda-a0be-b732304e11e9 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "interface-807d3025-d6a7-4778-a829-a61e2c7495c9-62c1052e-dc0a-4fb6-a399-3e238bf93789" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.537s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.343747] env[62569]: DEBUG nova.compute.manager [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 877.431806] env[62569]: DEBUG nova.network.neutron [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Updating instance_info_cache with network_info: [{"id": "19f8c74e-ebd3-4b14-9e7b-1539ce706834", "address": "fa:16:3e:6c:81:0e", "network": {"id": "30d3bf4a-8f55-492f-a049-181e29731818", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1244949475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bdba6022e3d4697a336ca28ca4eccec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19f8c74e-eb", "ovs_interfaceid": "19f8c74e-ebd3-4b14-9e7b-1539ce706834", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.630462] env[62569]: DEBUG oslo_concurrency.lockutils [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.309s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.630462] env[62569]: INFO nova.compute.manager [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Migrating [ 877.630462] env[62569]: DEBUG oslo_concurrency.lockutils [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.630678] env[62569]: DEBUG oslo_concurrency.lockutils [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired lock "compute-rpcapi-router" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.635023] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.426s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.636901] env[62569]: INFO nova.compute.claims [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 877.732804] env[62569]: DEBUG oslo_vmware.rw_handles [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527173ad-8bf5-4446-8686-6db5d632a24f/disk-0.vmdk. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 877.734065] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6232c11d-ef5e-4211-9932-c1a8207eb986 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.744578] env[62569]: DEBUG oslo_vmware.rw_handles [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527173ad-8bf5-4446-8686-6db5d632a24f/disk-0.vmdk is in state: ready. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 877.744578] env[62569]: ERROR oslo_vmware.rw_handles [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527173ad-8bf5-4446-8686-6db5d632a24f/disk-0.vmdk due to incomplete transfer. [ 877.744863] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-a11ffdff-add7-4af9-8162-b8199e4bfd72 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.754110] env[62569]: DEBUG oslo_vmware.rw_handles [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527173ad-8bf5-4446-8686-6db5d632a24f/disk-0.vmdk. {{(pid=62569) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 877.754110] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Uploaded image 870dec0b-8c00-43e0-870e-316f6bd64e1e to the Glance image server {{(pid=62569) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 877.756426] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Destroying the VM {{(pid=62569) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 877.756740] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-183b97bd-fc3e-4d77-ba61-25aec804c636 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.767010] env[62569]: DEBUG oslo_vmware.api [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 877.767010] env[62569]: value = "task-1250262" [ 877.767010] env[62569]: _type = "Task" [ 877.767010] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.782364] env[62569]: DEBUG oslo_vmware.api [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250262, 'name': Destroy_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.789479] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250261, 'name': MoveVirtualDisk_Task} progress is 29%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.866589] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.933589] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Releasing lock "refresh_cache-108d5bde-e463-4a69-816b-bdd3c03e13cd" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.934398] env[62569]: DEBUG nova.compute.manager [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Instance network_info: |[{"id": "19f8c74e-ebd3-4b14-9e7b-1539ce706834", "address": "fa:16:3e:6c:81:0e", "network": {"id": "30d3bf4a-8f55-492f-a049-181e29731818", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1244949475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bdba6022e3d4697a336ca28ca4eccec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19f8c74e-eb", "ovs_interfaceid": "19f8c74e-ebd3-4b14-9e7b-1539ce706834", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 877.934977] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:81:0e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '983826cf-6390-4ec6-bf97-30a1060947fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '19f8c74e-ebd3-4b14-9e7b-1539ce706834', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 877.944559] env[62569]: DEBUG oslo.service.loopingcall [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 877.944879] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 877.945211] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-02d774a7-38c8-477e-b1c1-85fc17b7595d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.970743] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 877.970743] env[62569]: value = "task-1250263" [ 877.970743] env[62569]: _type = "Task" [ 877.970743] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.982340] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250263, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.053736] env[62569]: DEBUG nova.objects.instance [None req-526c80e3-116e-40ad-9699-bd50a178ad19 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lazy-loading 'flavor' on Instance uuid e4eadcdf-a04a-4255-ba1c-fe20156c655f {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 878.142269] env[62569]: INFO nova.compute.rpcapi [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Automatically selected compute RPC version 6.4 from minimum service version 68 [ 878.142269] env[62569]: DEBUG oslo_concurrency.lockutils [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Releasing lock "compute-rpcapi-router" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.189242] env[62569]: DEBUG oslo_concurrency.lockutils [None req-37362a0d-b94c-4a40-9618-c0dec3e7b366 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Acquiring lock "298efba9-226d-4105-8b32-76b3dddddb0b" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.189584] env[62569]: DEBUG oslo_concurrency.lockutils [None req-37362a0d-b94c-4a40-9618-c0dec3e7b366 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Lock "298efba9-226d-4105-8b32-76b3dddddb0b" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.189856] env[62569]: INFO nova.compute.manager [None req-37362a0d-b94c-4a40-9618-c0dec3e7b366 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Rebooting instance [ 878.283996] env[62569]: DEBUG oslo_vmware.api [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250262, 'name': Destroy_Task, 'duration_secs': 0.425311} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.284362] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Destroyed the VM [ 878.284780] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Deleting Snapshot of the VM instance {{(pid=62569) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 878.285647] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-697672da-b18e-4fa3-ae6f-b30047d35c3a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.294022] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250261, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.295815] env[62569]: DEBUG oslo_vmware.api [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 878.295815] env[62569]: value = "task-1250264" [ 878.295815] env[62569]: _type = "Task" [ 878.295815] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.307399] env[62569]: DEBUG oslo_vmware.api [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250264, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.483994] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250263, 'name': CreateVM_Task, 'duration_secs': 0.444392} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.485038] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 878.486227] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.486410] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.487030] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 878.487360] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eeeaa125-3e2f-4516-b7a8-ed723b3bf6c4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.501041] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 878.501041] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b8f220-0d9d-1af0-07fa-74762ef8573b" [ 878.501041] env[62569]: _type = "Task" [ 878.501041] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.514239] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b8f220-0d9d-1af0-07fa-74762ef8573b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.559679] env[62569]: DEBUG oslo_concurrency.lockutils [None req-526c80e3-116e-40ad-9699-bd50a178ad19 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.559812] env[62569]: DEBUG oslo_concurrency.lockutils [None req-526c80e3-116e-40ad-9699-bd50a178ad19 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.559992] env[62569]: DEBUG nova.network.neutron [None req-526c80e3-116e-40ad-9699-bd50a178ad19 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 878.560251] env[62569]: DEBUG nova.objects.instance [None req-526c80e3-116e-40ad-9699-bd50a178ad19 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lazy-loading 'info_cache' on Instance uuid e4eadcdf-a04a-4255-ba1c-fe20156c655f {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 878.669047] env[62569]: DEBUG oslo_concurrency.lockutils [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "refresh_cache-31ce29fa-4f60-4404-b830-21ad196f78b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.669047] env[62569]: DEBUG oslo_concurrency.lockutils [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired lock "refresh_cache-31ce29fa-4f60-4404-b830-21ad196f78b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.669047] env[62569]: DEBUG nova.network.neutron [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 878.723204] env[62569]: DEBUG oslo_concurrency.lockutils [None req-37362a0d-b94c-4a40-9618-c0dec3e7b366 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Acquiring lock "refresh_cache-298efba9-226d-4105-8b32-76b3dddddb0b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.723204] env[62569]: DEBUG oslo_concurrency.lockutils [None req-37362a0d-b94c-4a40-9618-c0dec3e7b366 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Acquired lock "refresh_cache-298efba9-226d-4105-8b32-76b3dddddb0b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.723349] env[62569]: DEBUG nova.network.neutron [None req-37362a0d-b94c-4a40-9618-c0dec3e7b366 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 878.795133] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250261, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.809839] env[62569]: DEBUG oslo_vmware.api [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250264, 'name': RemoveSnapshot_Task} progress is 36%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.929421] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b51303e-f4d2-4f16-8c52-33aaa4ae71bf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.939029] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1554657-94b1-4f81-a163-55d7d5f355b2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.973604] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a67e99d-81c2-4366-8044-94c1410df7cd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.983877] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a529b8f-1d2a-4703-8508-aec5be15e2a2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.010145] env[62569]: DEBUG nova.compute.provider_tree [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.030557] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b8f220-0d9d-1af0-07fa-74762ef8573b, 'name': SearchDatastore_Task, 'duration_secs': 0.07987} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.031197] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.031653] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 879.032109] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.032431] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.032802] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 879.033286] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f8ea69ab-9988-4c5c-8eaa-a5a9f106095b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.054011] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 879.054260] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 879.055067] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-adc2dbfd-d1ce-4fbb-848c-8861806d17e8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.063539] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 879.063539] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52fbb486-3ef5-f61b-ce21-fdf7fbd9e954" [ 879.063539] env[62569]: _type = "Task" [ 879.063539] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.065027] env[62569]: DEBUG nova.objects.base [None req-526c80e3-116e-40ad-9699-bd50a178ad19 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62569) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 879.077878] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52fbb486-3ef5-f61b-ce21-fdf7fbd9e954, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.295290] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250261, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.313255] env[62569]: DEBUG oslo_vmware.api [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250264, 'name': RemoveSnapshot_Task, 'duration_secs': 0.77127} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.313602] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Deleted Snapshot of the VM instance {{(pid=62569) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 879.313828] env[62569]: INFO nova.compute.manager [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Took 15.35 seconds to snapshot the instance on the hypervisor. [ 879.396147] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.396440] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.476414] env[62569]: DEBUG nova.network.neutron [None req-37362a0d-b94c-4a40-9618-c0dec3e7b366 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Updating instance_info_cache with network_info: [{"id": "f33064d8-b284-4ed3-9a2b-01f27371be83", "address": "fa:16:3e:0b:05:c4", "network": {"id": "8a82d478-4aa8-4179-b974-51200b4e2bc6", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1314992403-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1af23d0979b84b85b348b3868379ef2b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf33064d8-b2", "ovs_interfaceid": "f33064d8-b284-4ed3-9a2b-01f27371be83", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.519991] env[62569]: DEBUG nova.scheduler.client.report [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 879.574845] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52fbb486-3ef5-f61b-ce21-fdf7fbd9e954, 'name': SearchDatastore_Task, 'duration_secs': 0.093305} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.575960] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8af9fa94-5d70-4a72-a9fb-82381ce953bd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.580950] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 879.580950] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52603ae4-c0a5-e6aa-046b-5c2e95913fde" [ 879.580950] env[62569]: _type = "Task" [ 879.580950] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.588259] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52603ae4-c0a5-e6aa-046b-5c2e95913fde, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.696657] env[62569]: DEBUG nova.compute.manager [req-cf26eac1-2e3f-4ff3-b402-6b2bd63b1296 req-3e524f2f-0bfa-4f65-a6f9-eed34b178ee8 service nova] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Received event network-changed-19f8c74e-ebd3-4b14-9e7b-1539ce706834 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 879.696765] env[62569]: DEBUG nova.compute.manager [req-cf26eac1-2e3f-4ff3-b402-6b2bd63b1296 req-3e524f2f-0bfa-4f65-a6f9-eed34b178ee8 service nova] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Refreshing instance network info cache due to event network-changed-19f8c74e-ebd3-4b14-9e7b-1539ce706834. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 879.696957] env[62569]: DEBUG oslo_concurrency.lockutils [req-cf26eac1-2e3f-4ff3-b402-6b2bd63b1296 req-3e524f2f-0bfa-4f65-a6f9-eed34b178ee8 service nova] Acquiring lock "refresh_cache-108d5bde-e463-4a69-816b-bdd3c03e13cd" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.697121] env[62569]: DEBUG oslo_concurrency.lockutils [req-cf26eac1-2e3f-4ff3-b402-6b2bd63b1296 req-3e524f2f-0bfa-4f65-a6f9-eed34b178ee8 service nova] Acquired lock "refresh_cache-108d5bde-e463-4a69-816b-bdd3c03e13cd" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.697285] env[62569]: DEBUG nova.network.neutron [req-cf26eac1-2e3f-4ff3-b402-6b2bd63b1296 req-3e524f2f-0bfa-4f65-a6f9-eed34b178ee8 service nova] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Refreshing network info cache for port 19f8c74e-ebd3-4b14-9e7b-1539ce706834 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 879.794290] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250261, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.751711} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.794570] env[62569]: INFO nova.virt.vmwareapi.ds_util [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_ee1abb3b-b3e0-49fa-8bb4-04290a98ab3c/OSTACK_IMG_ee1abb3b-b3e0-49fa-8bb4-04290a98ab3c.vmdk to [datastore2] devstack-image-cache_base/45f3d90c-b69b-4905-993e-439633bfeb29/45f3d90c-b69b-4905-993e-439633bfeb29.vmdk. [ 879.794758] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Cleaning up location [datastore2] OSTACK_IMG_ee1abb3b-b3e0-49fa-8bb4-04290a98ab3c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 879.794922] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_ee1abb3b-b3e0-49fa-8bb4-04290a98ab3c {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 879.795434] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6b9da440-a329-4fce-a277-c13720435470 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.802074] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 879.802074] env[62569]: value = "task-1250265" [ 879.802074] env[62569]: _type = "Task" [ 879.802074] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.806169] env[62569]: DEBUG nova.network.neutron [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Updating instance_info_cache with network_info: [{"id": "8e55dca6-f777-4333-9b49-feb5527d9a5e", "address": "fa:16:3e:97:39:72", "network": {"id": "c61b7867-7897-43ae-8f5b-19834c4390ea", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-512066189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ba7e96d226942bd99ab29fe703c6fea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e55dca6-f7", "ovs_interfaceid": "8e55dca6-f777-4333-9b49-feb5527d9a5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.814997] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250265, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.823782] env[62569]: DEBUG nova.network.neutron [None req-526c80e3-116e-40ad-9699-bd50a178ad19 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Updating instance_info_cache with network_info: [{"id": "39e71a4f-3b81-4b34-8cc2-2a5352eca4ce", "address": "fa:16:3e:4b:e2:a8", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap39e71a4f-3b", "ovs_interfaceid": "39e71a4f-3b81-4b34-8cc2-2a5352eca4ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.870271] env[62569]: DEBUG nova.compute.manager [None req-e59c62ac-3968-4140-a9e4-66ae23c53840 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Found 1 images (rotation: 2) {{(pid=62569) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4923}} [ 879.901189] env[62569]: DEBUG nova.compute.manager [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 879.980226] env[62569]: DEBUG oslo_concurrency.lockutils [None req-37362a0d-b94c-4a40-9618-c0dec3e7b366 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Releasing lock "refresh_cache-298efba9-226d-4105-8b32-76b3dddddb0b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.028018] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.394s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.028018] env[62569]: DEBUG nova.compute.manager [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 880.029900] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.295s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.031599] env[62569]: INFO nova.compute.claims [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 880.092771] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52603ae4-c0a5-e6aa-046b-5c2e95913fde, 'name': SearchDatastore_Task, 'duration_secs': 0.010037} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.093055] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.093359] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 108d5bde-e463-4a69-816b-bdd3c03e13cd/108d5bde-e463-4a69-816b-bdd3c03e13cd.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 880.093651] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-639fc57e-3cbb-4068-8a83-24a0425617ad {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.099907] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 880.099907] env[62569]: value = "task-1250266" [ 880.099907] env[62569]: _type = "Task" [ 880.099907] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.107440] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250266, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.310356] env[62569]: DEBUG oslo_concurrency.lockutils [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Releasing lock "refresh_cache-31ce29fa-4f60-4404-b830-21ad196f78b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.314194] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250265, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.035403} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.314606] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 880.314682] env[62569]: DEBUG oslo_concurrency.lockutils [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Releasing lock "[datastore2] devstack-image-cache_base/45f3d90c-b69b-4905-993e-439633bfeb29/45f3d90c-b69b-4905-993e-439633bfeb29.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.314885] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/45f3d90c-b69b-4905-993e-439633bfeb29/45f3d90c-b69b-4905-993e-439633bfeb29.vmdk to [datastore2] 337d88a2-30b9-4846-929e-042bd7a64a65/337d88a2-30b9-4846-929e-042bd7a64a65.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 880.315184] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-36ac81d2-4ec2-4b23-a74b-2e753907fc97 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.325685] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 880.325685] env[62569]: value = "task-1250267" [ 880.325685] env[62569]: _type = "Task" [ 880.325685] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.329860] env[62569]: DEBUG oslo_concurrency.lockutils [None req-526c80e3-116e-40ad-9699-bd50a178ad19 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.343311] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250267, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.350480] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Acquiring lock "691d8fbe-b9ee-454f-bd7b-14520e53ed26" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.350659] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Lock "691d8fbe-b9ee-454f-bd7b-14520e53ed26" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.351964] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Acquiring lock "691d8fbe-b9ee-454f-bd7b-14520e53ed26-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.351964] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Lock "691d8fbe-b9ee-454f-bd7b-14520e53ed26-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.351964] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Lock "691d8fbe-b9ee-454f-bd7b-14520e53ed26-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.354676] env[62569]: INFO nova.compute.manager [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Terminating instance [ 880.431024] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.453065] env[62569]: DEBUG nova.network.neutron [req-cf26eac1-2e3f-4ff3-b402-6b2bd63b1296 req-3e524f2f-0bfa-4f65-a6f9-eed34b178ee8 service nova] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Updated VIF entry in instance network info cache for port 19f8c74e-ebd3-4b14-9e7b-1539ce706834. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 880.453493] env[62569]: DEBUG nova.network.neutron [req-cf26eac1-2e3f-4ff3-b402-6b2bd63b1296 req-3e524f2f-0bfa-4f65-a6f9-eed34b178ee8 service nova] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Updating instance_info_cache with network_info: [{"id": "19f8c74e-ebd3-4b14-9e7b-1539ce706834", "address": "fa:16:3e:6c:81:0e", "network": {"id": "30d3bf4a-8f55-492f-a049-181e29731818", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1244949475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bdba6022e3d4697a336ca28ca4eccec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19f8c74e-eb", "ovs_interfaceid": "19f8c74e-ebd3-4b14-9e7b-1539ce706834", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.486555] env[62569]: DEBUG nova.compute.manager [None req-37362a0d-b94c-4a40-9618-c0dec3e7b366 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 880.487597] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67f0c324-5723-4af9-9f21-f66663828c6c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.531786] env[62569]: DEBUG nova.compute.utils [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 880.533913] env[62569]: DEBUG nova.compute.manager [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 880.534242] env[62569]: DEBUG nova.network.neutron [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 880.557065] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "interface-807d3025-d6a7-4778-a829-a61e2c7495c9-5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.557346] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "interface-807d3025-d6a7-4778-a829-a61e2c7495c9-5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.586670] env[62569]: DEBUG nova.policy [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a272992a4894805bfb958680a37da8e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b20340a1ce0447cae1bfd8b7c28928f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 880.613048] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250266, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473514} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.613048] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 108d5bde-e463-4a69-816b-bdd3c03e13cd/108d5bde-e463-4a69-816b-bdd3c03e13cd.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 880.613048] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 880.613048] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cc5e7ea9-2c81-4a07-b3c0-7287243308e7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.618100] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 880.618100] env[62569]: value = "task-1250268" [ 880.618100] env[62569]: _type = "Task" [ 880.618100] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.627065] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250268, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.845332] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250267, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.864793] env[62569]: DEBUG nova.compute.manager [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 880.864793] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 880.868017] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-385e56ff-c557-40eb-8fbe-56cdabd52d4c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.877347] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 880.878704] env[62569]: DEBUG nova.network.neutron [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Successfully created port: 6ca336c0-76d8-4913-91e7-f7c65a3aa5c1 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 880.881188] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b129188c-2197-421a-a5ac-244d080e1f0d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.896403] env[62569]: DEBUG oslo_vmware.api [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Waiting for the task: (returnval){ [ 880.896403] env[62569]: value = "task-1250269" [ 880.896403] env[62569]: _type = "Task" [ 880.896403] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.913648] env[62569]: DEBUG oslo_vmware.api [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Task: {'id': task-1250269, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.956417] env[62569]: DEBUG oslo_concurrency.lockutils [req-cf26eac1-2e3f-4ff3-b402-6b2bd63b1296 req-3e524f2f-0bfa-4f65-a6f9-eed34b178ee8 service nova] Releasing lock "refresh_cache-108d5bde-e463-4a69-816b-bdd3c03e13cd" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.048043] env[62569]: DEBUG nova.compute.manager [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 881.060441] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.061815] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.062271] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-155280ae-4100-4e80-8832-932285b583c7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.092428] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34901931-064e-42ab-bdc5-f991b7f8d03d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.138194] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Reconfiguring VM to detach interface {{(pid=62569) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 881.146370] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68efb750-8397-48d1-819b-510fc9eaf8f4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.174307] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250268, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.127415} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.174307] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 881.174650] env[62569]: DEBUG oslo_vmware.api [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 881.174650] env[62569]: value = "task-1250270" [ 881.174650] env[62569]: _type = "Task" [ 881.174650] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.175365] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a47c00e-6163-43ec-b48a-577c28327c36 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.194135] env[62569]: DEBUG oslo_vmware.api [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250270, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.213947] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 108d5bde-e463-4a69-816b-bdd3c03e13cd/108d5bde-e463-4a69-816b-bdd3c03e13cd.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 881.217032] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a4e27bc-fe02-4fa5-8a55-27467d71f8fd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.239608] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 881.239608] env[62569]: value = "task-1250271" [ 881.239608] env[62569]: _type = "Task" [ 881.239608] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.254353] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250271, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.343515] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250267, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.347618] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-526c80e3-116e-40ad-9699-bd50a178ad19 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 881.347955] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7bec931d-956e-4acc-ba8e-2c36b8f74278 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.357122] env[62569]: DEBUG oslo_vmware.api [None req-526c80e3-116e-40ad-9699-bd50a178ad19 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 881.357122] env[62569]: value = "task-1250272" [ 881.357122] env[62569]: _type = "Task" [ 881.357122] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.369948] env[62569]: DEBUG oslo_vmware.api [None req-526c80e3-116e-40ad-9699-bd50a178ad19 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250272, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.391190] env[62569]: DEBUG nova.compute.manager [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 881.392207] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc9f1ec9-c305-4bd8-8da0-afb5b3d68b4a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.413326] env[62569]: DEBUG oslo_vmware.api [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Task: {'id': task-1250269, 'name': PowerOffVM_Task, 'duration_secs': 0.207851} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.413663] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 881.413838] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 881.414157] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-01b1f4c4-4bc3-41b4-b6f8-ec47a4c09f18 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.466723] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ea3f214-f0bc-4fd8-8c22-1f77f2c7dfeb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.476068] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49280ca8-c067-4b36-b251-28ef096f975d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.485332] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 881.485729] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 881.485932] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Deleting the datastore file [datastore1] 691d8fbe-b9ee-454f-bd7b-14520e53ed26 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 881.486723] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-30dd11a3-8d95-4983-a369-3fdd0ae0baaa {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.520554] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26707b88-3c7e-49e1-b132-609f7408d8c2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.524805] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9d3d6b0-d3f8-402a-b85d-d46c03c49e45 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.528041] env[62569]: DEBUG oslo_vmware.api [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Waiting for the task: (returnval){ [ 881.528041] env[62569]: value = "task-1250274" [ 881.528041] env[62569]: _type = "Task" [ 881.528041] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.535546] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-37362a0d-b94c-4a40-9618-c0dec3e7b366 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Doing hard reboot of VM {{(pid=62569) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 881.541015] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-b7c264fb-75e8-49fd-951a-04f53850420a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.542504] env[62569]: DEBUG oslo_vmware.api [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Task: {'id': task-1250274, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.544429] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3e9fcf8-3866-4a48-a6e0-b6f94a47e04c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.565995] env[62569]: DEBUG nova.compute.provider_tree [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.571027] env[62569]: DEBUG oslo_vmware.api [None req-37362a0d-b94c-4a40-9618-c0dec3e7b366 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Waiting for the task: (returnval){ [ 881.571027] env[62569]: value = "task-1250275" [ 881.571027] env[62569]: _type = "Task" [ 881.571027] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.581323] env[62569]: DEBUG oslo_vmware.api [None req-37362a0d-b94c-4a40-9618-c0dec3e7b366 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Task: {'id': task-1250275, 'name': ResetVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.690200] env[62569]: DEBUG oslo_vmware.api [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250270, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.752997] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250271, 'name': ReconfigVM_Task, 'duration_secs': 0.488585} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.753625] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 108d5bde-e463-4a69-816b-bdd3c03e13cd/108d5bde-e463-4a69-816b-bdd3c03e13cd.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 881.754663] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cbc268eb-c666-4959-ba35-db06eadf75fb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.766020] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 881.766020] env[62569]: value = "task-1250276" [ 881.766020] env[62569]: _type = "Task" [ 881.766020] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.775308] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250276, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.833169] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8307afc-6aa1-4bf5-8952-03ea012dd0a2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.856326] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250267, 'name': CopyVirtualDisk_Task} progress is 54%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.870238] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Updating instance '31ce29fa-4f60-4404-b830-21ad196f78b5' progress to 0 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 881.885993] env[62569]: DEBUG oslo_vmware.api [None req-526c80e3-116e-40ad-9699-bd50a178ad19 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250272, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.911186] env[62569]: INFO nova.compute.manager [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] instance snapshotting [ 881.911186] env[62569]: DEBUG nova.objects.instance [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lazy-loading 'flavor' on Instance uuid b47cd2d7-0cd2-41af-8ed1-a6dfca323516 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 882.040336] env[62569]: DEBUG oslo_vmware.api [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Task: {'id': task-1250274, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.23889} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.041320] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 882.041668] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 882.041987] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 882.042286] env[62569]: INFO nova.compute.manager [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Took 1.18 seconds to destroy the instance on the hypervisor. [ 882.042643] env[62569]: DEBUG oslo.service.loopingcall [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.042924] env[62569]: DEBUG nova.compute.manager [-] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 882.043087] env[62569]: DEBUG nova.network.neutron [-] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 882.070251] env[62569]: DEBUG nova.compute.manager [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 882.078269] env[62569]: DEBUG nova.scheduler.client.report [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 882.090714] env[62569]: DEBUG oslo_vmware.api [None req-37362a0d-b94c-4a40-9618-c0dec3e7b366 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Task: {'id': task-1250275, 'name': ResetVM_Task, 'duration_secs': 0.12918} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.093428] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-37362a0d-b94c-4a40-9618-c0dec3e7b366 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Did hard reboot of VM {{(pid=62569) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 882.093428] env[62569]: DEBUG nova.compute.manager [None req-37362a0d-b94c-4a40-9618-c0dec3e7b366 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 882.093428] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7751b68b-296d-4f9b-b78b-57073353b6cc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.108836] env[62569]: DEBUG nova.virt.hardware [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='2fdb8a55f13ae538cae4e9635e8ffdb8',container_format='bare',created_at=2024-10-10T15:37:38Z,direct_url=,disk_format='vmdk',id=5909e6a6-6c00-4956-a4d0-c6430d91b165,min_disk=1,min_ram=0,name='tempest-test-snap-451123963',owner='7b20340a1ce0447cae1bfd8b7c28928f',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-10T15:37:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 882.108836] env[62569]: DEBUG nova.virt.hardware [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 882.108836] env[62569]: DEBUG nova.virt.hardware [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 882.109127] env[62569]: DEBUG nova.virt.hardware [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 882.109227] env[62569]: DEBUG nova.virt.hardware [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 882.109424] env[62569]: DEBUG nova.virt.hardware [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 882.109689] env[62569]: DEBUG nova.virt.hardware [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 882.109894] env[62569]: DEBUG nova.virt.hardware [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 882.110641] env[62569]: DEBUG nova.virt.hardware [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 882.110641] env[62569]: DEBUG nova.virt.hardware [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 882.110641] env[62569]: DEBUG nova.virt.hardware [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 882.111724] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6028c6ad-2577-4883-b496-f418c4a7e214 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.121630] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4771eacb-26f1-482c-aed6-0f5e56b4fb0c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.191291] env[62569]: DEBUG oslo_vmware.api [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250270, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.277349] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250276, 'name': Rename_Task, 'duration_secs': 0.260328} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.277349] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 882.278470] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-421c3673-a114-4945-9733-3e20abb59e5f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.289887] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 882.289887] env[62569]: value = "task-1250277" [ 882.289887] env[62569]: _type = "Task" [ 882.289887] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.300101] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250277, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.345383] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250267, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.384770] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 882.385303] env[62569]: DEBUG oslo_vmware.api [None req-526c80e3-116e-40ad-9699-bd50a178ad19 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250272, 'name': PowerOnVM_Task, 'duration_secs': 0.612161} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.385622] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aee453c6-581d-42b7-9537-c11e0240468a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.387497] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-526c80e3-116e-40ad-9699-bd50a178ad19 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 882.387781] env[62569]: DEBUG nova.compute.manager [None req-526c80e3-116e-40ad-9699-bd50a178ad19 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 882.388654] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2686b8-512e-4e3f-af4b-83714eb83949 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.404012] env[62569]: DEBUG oslo_vmware.api [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 882.404012] env[62569]: value = "task-1250278" [ 882.404012] env[62569]: _type = "Task" [ 882.404012] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.418969] env[62569]: DEBUG oslo_vmware.api [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250278, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.420989] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a35d81a-5b95-43ed-8f1c-9ae80cd9a1f3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.451589] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80029867-943f-4021-8780-8ef87389ba03 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.461143] env[62569]: DEBUG nova.compute.manager [req-bde9f2e4-b186-48ca-a4b7-9d23f0dac4cc req-b1c7c11c-f416-46ae-94cb-4bde085d4279 service nova] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Received event network-vif-plugged-6ca336c0-76d8-4913-91e7-f7c65a3aa5c1 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 882.461398] env[62569]: DEBUG oslo_concurrency.lockutils [req-bde9f2e4-b186-48ca-a4b7-9d23f0dac4cc req-b1c7c11c-f416-46ae-94cb-4bde085d4279 service nova] Acquiring lock "9c1c0379-3968-464b-b587-8fac704404d5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.461614] env[62569]: DEBUG oslo_concurrency.lockutils [req-bde9f2e4-b186-48ca-a4b7-9d23f0dac4cc req-b1c7c11c-f416-46ae-94cb-4bde085d4279 service nova] Lock "9c1c0379-3968-464b-b587-8fac704404d5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.461789] env[62569]: DEBUG oslo_concurrency.lockutils [req-bde9f2e4-b186-48ca-a4b7-9d23f0dac4cc req-b1c7c11c-f416-46ae-94cb-4bde085d4279 service nova] Lock "9c1c0379-3968-464b-b587-8fac704404d5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.461969] env[62569]: DEBUG nova.compute.manager [req-bde9f2e4-b186-48ca-a4b7-9d23f0dac4cc req-b1c7c11c-f416-46ae-94cb-4bde085d4279 service nova] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] No waiting events found dispatching network-vif-plugged-6ca336c0-76d8-4913-91e7-f7c65a3aa5c1 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 882.462456] env[62569]: WARNING nova.compute.manager [req-bde9f2e4-b186-48ca-a4b7-9d23f0dac4cc req-b1c7c11c-f416-46ae-94cb-4bde085d4279 service nova] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Received unexpected event network-vif-plugged-6ca336c0-76d8-4913-91e7-f7c65a3aa5c1 for instance with vm_state building and task_state spawning. [ 882.512123] env[62569]: DEBUG nova.network.neutron [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Successfully updated port: 6ca336c0-76d8-4913-91e7-f7c65a3aa5c1 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 882.585730] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.556s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.586416] env[62569]: DEBUG nova.compute.manager [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 882.589604] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.723s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.591601] env[62569]: INFO nova.compute.claims [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 882.619907] env[62569]: DEBUG oslo_concurrency.lockutils [None req-37362a0d-b94c-4a40-9618-c0dec3e7b366 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Lock "298efba9-226d-4105-8b32-76b3dddddb0b" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.430s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.692340] env[62569]: DEBUG oslo_vmware.api [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250270, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.803600] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250277, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.810879] env[62569]: DEBUG nova.compute.manager [req-a15b5300-fdd8-438c-b51c-5cb180fc3c2a req-97901357-8eb7-480d-858a-90a272b84cfa service nova] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Received event network-vif-deleted-0edf3387-b21d-47f6-b9e7-503d285ba554 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 882.811663] env[62569]: INFO nova.compute.manager [req-a15b5300-fdd8-438c-b51c-5cb180fc3c2a req-97901357-8eb7-480d-858a-90a272b84cfa service nova] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Neutron deleted interface 0edf3387-b21d-47f6-b9e7-503d285ba554; detaching it from the instance and deleting it from the info cache [ 882.811663] env[62569]: DEBUG nova.network.neutron [req-a15b5300-fdd8-438c-b51c-5cb180fc3c2a req-97901357-8eb7-480d-858a-90a272b84cfa service nova] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.847551] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250267, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.924539] env[62569]: DEBUG oslo_vmware.api [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250278, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.970564] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Creating Snapshot of the VM instance {{(pid=62569) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 882.972758] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9a469021-5a0b-471d-91f6-c60c5959a6db {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.979785] env[62569]: DEBUG oslo_vmware.api [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 882.979785] env[62569]: value = "task-1250279" [ 882.979785] env[62569]: _type = "Task" [ 882.979785] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.988965] env[62569]: DEBUG oslo_vmware.api [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250279, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.012609] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "refresh_cache-9c1c0379-3968-464b-b587-8fac704404d5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.012609] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquired lock "refresh_cache-9c1c0379-3968-464b-b587-8fac704404d5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.012609] env[62569]: DEBUG nova.network.neutron [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 883.055897] env[62569]: DEBUG nova.network.neutron [-] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.095637] env[62569]: DEBUG nova.compute.utils [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 883.102231] env[62569]: DEBUG nova.compute.manager [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 883.102231] env[62569]: DEBUG nova.network.neutron [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 883.171825] env[62569]: DEBUG nova.policy [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '574a84c78a2b440fb69e2f102f8f8a0f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '335bc4885a9e4675a3357199a612faed', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 883.190565] env[62569]: DEBUG oslo_vmware.api [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250270, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.302405] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250277, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.316587] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ba0598b1-08d7-42a6-ae72-18f97d32e231 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.327021] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbf465ed-0f07-4dee-927a-994a09239e87 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.342804] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250267, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.581064} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.343467] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/45f3d90c-b69b-4905-993e-439633bfeb29/45f3d90c-b69b-4905-993e-439633bfeb29.vmdk to [datastore2] 337d88a2-30b9-4846-929e-042bd7a64a65/337d88a2-30b9-4846-929e-042bd7a64a65.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 883.344474] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20e5ffba-e458-4075-9408-0e6e9ae57db9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.357575] env[62569]: DEBUG nova.compute.manager [req-a15b5300-fdd8-438c-b51c-5cb180fc3c2a req-97901357-8eb7-480d-858a-90a272b84cfa service nova] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Detach interface failed, port_id=0edf3387-b21d-47f6-b9e7-503d285ba554, reason: Instance 691d8fbe-b9ee-454f-bd7b-14520e53ed26 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 883.388375] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 337d88a2-30b9-4846-929e-042bd7a64a65/337d88a2-30b9-4846-929e-042bd7a64a65.vmdk or device None with type streamOptimized {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 883.389779] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-abd8a280-2806-4c04-87fc-9655c74867ef {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.432480] env[62569]: DEBUG oslo_vmware.api [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250278, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.434136] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 883.434136] env[62569]: value = "task-1250280" [ 883.434136] env[62569]: _type = "Task" [ 883.434136] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.442262] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250280, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.490830] env[62569]: DEBUG oslo_vmware.api [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250279, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.548804] env[62569]: DEBUG nova.network.neutron [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 883.557623] env[62569]: INFO nova.compute.manager [-] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Took 1.51 seconds to deallocate network for instance. [ 883.603999] env[62569]: DEBUG nova.compute.manager [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 883.639329] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Acquiring lock "298efba9-226d-4105-8b32-76b3dddddb0b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.639688] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Lock "298efba9-226d-4105-8b32-76b3dddddb0b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.639942] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Acquiring lock "298efba9-226d-4105-8b32-76b3dddddb0b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.640258] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Lock "298efba9-226d-4105-8b32-76b3dddddb0b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.640446] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Lock "298efba9-226d-4105-8b32-76b3dddddb0b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.646289] env[62569]: INFO nova.compute.manager [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Terminating instance [ 883.695536] env[62569]: DEBUG oslo_vmware.api [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250270, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.739391] env[62569]: DEBUG nova.network.neutron [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Updating instance_info_cache with network_info: [{"id": "6ca336c0-76d8-4913-91e7-f7c65a3aa5c1", "address": "fa:16:3e:b8:41:ba", "network": {"id": "8ea5e138-813b-4c5c-88a1-b1ea79807f10", "bridge": "br-int", "label": "tempest-ImagesTestJSON-524163445-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b20340a1ce0447cae1bfd8b7c28928f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ca336c0-76", "ovs_interfaceid": "6ca336c0-76d8-4913-91e7-f7c65a3aa5c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.774180] env[62569]: DEBUG nova.network.neutron [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Successfully created port: 057050a7-da70-4b2f-bd56-b5de3a3dbb31 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 883.804985] env[62569]: DEBUG oslo_vmware.api [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250277, 'name': PowerOnVM_Task, 'duration_secs': 1.338148} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.806503] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 883.806503] env[62569]: INFO nova.compute.manager [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Took 8.36 seconds to spawn the instance on the hypervisor. [ 883.806503] env[62569]: DEBUG nova.compute.manager [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 883.806706] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a022e6-e1f3-4307-aa4b-36b7a3321db8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.916160] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac75bce7-6c03-4240-8341-a85f880d6bb6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.926847] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-015500e6-5080-4e76-ac5f-cb072d08a943 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.936094] env[62569]: DEBUG oslo_vmware.api [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250278, 'name': PowerOffVM_Task, 'duration_secs': 1.323142} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.939725] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 883.940192] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Updating instance '31ce29fa-4f60-4404-b830-21ad196f78b5' progress to 17 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 883.978567] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20b9c3ad-92ff-47a1-a60b-87ee68d9832f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.985180] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250280, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.996229] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73ffbaf8-ce46-454c-9e65-d5a49d3b01dd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.000290] env[62569]: DEBUG oslo_vmware.api [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250279, 'name': CreateSnapshot_Task, 'duration_secs': 0.618054} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.001408] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Created Snapshot of the VM instance {{(pid=62569) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 884.002671] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68f4d432-df71-4e83-8e58-cd256e0e01a0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.013470] env[62569]: DEBUG nova.compute.provider_tree [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.065223] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.154796] env[62569]: DEBUG nova.compute.manager [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 884.154796] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 884.154796] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec7144ae-9a42-4445-8070-93a3d1672501 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.162096] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 884.162421] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cac3507e-5669-486c-9cec-383fe4ed7eee {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.169557] env[62569]: DEBUG oslo_vmware.api [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Waiting for the task: (returnval){ [ 884.169557] env[62569]: value = "task-1250281" [ 884.169557] env[62569]: _type = "Task" [ 884.169557] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.177795] env[62569]: DEBUG oslo_vmware.api [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Task: {'id': task-1250281, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.191335] env[62569]: DEBUG oslo_vmware.api [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250270, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.244765] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Releasing lock "refresh_cache-9c1c0379-3968-464b-b587-8fac704404d5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.245139] env[62569]: DEBUG nova.compute.manager [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Instance network_info: |[{"id": "6ca336c0-76d8-4913-91e7-f7c65a3aa5c1", "address": "fa:16:3e:b8:41:ba", "network": {"id": "8ea5e138-813b-4c5c-88a1-b1ea79807f10", "bridge": "br-int", "label": "tempest-ImagesTestJSON-524163445-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b20340a1ce0447cae1bfd8b7c28928f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ca336c0-76", "ovs_interfaceid": "6ca336c0-76d8-4913-91e7-f7c65a3aa5c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 884.246032] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b8:41:ba', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f65996a3-f865-4492-9377-cd14ec8b3aae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6ca336c0-76d8-4913-91e7-f7c65a3aa5c1', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 884.254425] env[62569]: DEBUG oslo.service.loopingcall [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.256267] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 884.257229] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cc7fe3ed-2cbe-4bea-81e5-31c2da286354 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.278820] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 884.278820] env[62569]: value = "task-1250282" [ 884.278820] env[62569]: _type = "Task" [ 884.278820] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.288160] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250282, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.336448] env[62569]: INFO nova.compute.manager [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Took 19.48 seconds to build instance. [ 884.446367] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250280, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.480024] env[62569]: DEBUG nova.virt.hardware [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 884.480024] env[62569]: DEBUG nova.virt.hardware [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 884.480024] env[62569]: DEBUG nova.virt.hardware [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 884.484517] env[62569]: DEBUG nova.virt.hardware [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 884.484517] env[62569]: DEBUG nova.virt.hardware [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 884.484517] env[62569]: DEBUG nova.virt.hardware [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 884.484517] env[62569]: DEBUG nova.virt.hardware [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 884.484517] env[62569]: DEBUG nova.virt.hardware [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 884.484517] env[62569]: DEBUG nova.virt.hardware [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 884.484517] env[62569]: DEBUG nova.virt.hardware [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 884.484517] env[62569]: DEBUG nova.virt.hardware [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 884.488714] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-81df6ad1-087b-4eda-b5c9-c9cb727ac02b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.508522] env[62569]: DEBUG oslo_vmware.api [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 884.508522] env[62569]: value = "task-1250283" [ 884.508522] env[62569]: _type = "Task" [ 884.508522] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.519591] env[62569]: DEBUG nova.scheduler.client.report [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 884.532551] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Creating linked-clone VM from snapshot {{(pid=62569) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 884.533582] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-587dec7c-a9f7-4011-89e3-3edefde9dacc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.541845] env[62569]: DEBUG oslo_vmware.api [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250283, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.548053] env[62569]: DEBUG oslo_vmware.api [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 884.548053] env[62569]: value = "task-1250284" [ 884.548053] env[62569]: _type = "Task" [ 884.548053] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.557486] env[62569]: DEBUG oslo_vmware.api [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250284, 'name': CloneVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.616926] env[62569]: DEBUG nova.compute.manager [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 884.642072] env[62569]: DEBUG nova.virt.hardware [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 884.642491] env[62569]: DEBUG nova.virt.hardware [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 884.642571] env[62569]: DEBUG nova.virt.hardware [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 884.642793] env[62569]: DEBUG nova.virt.hardware [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 884.643075] env[62569]: DEBUG nova.virt.hardware [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 884.643137] env[62569]: DEBUG nova.virt.hardware [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 884.643395] env[62569]: DEBUG nova.virt.hardware [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 884.643602] env[62569]: DEBUG nova.virt.hardware [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 884.643831] env[62569]: DEBUG nova.virt.hardware [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 884.644043] env[62569]: DEBUG nova.virt.hardware [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 884.644286] env[62569]: DEBUG nova.virt.hardware [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 884.645339] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba9be0e2-0926-4f95-ba46-560e30fbe01f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.653745] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-569c7125-8e34-4831-a5a3-c72b235a6e65 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.678971] env[62569]: DEBUG oslo_vmware.api [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Task: {'id': task-1250281, 'name': PowerOffVM_Task, 'duration_secs': 0.182669} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.679288] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 884.679475] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 884.679721] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-60178a07-51cf-48af-9820-0daa8ca6b3f4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.691120] env[62569]: DEBUG oslo_vmware.api [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250270, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.728179] env[62569]: DEBUG nova.compute.manager [req-7f96f7f4-043a-45f1-ab8d-f2c67912da66 req-06ee71df-65c1-40c8-b460-68d2213a9315 service nova] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Received event network-changed-6ca336c0-76d8-4913-91e7-f7c65a3aa5c1 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 884.728179] env[62569]: DEBUG nova.compute.manager [req-7f96f7f4-043a-45f1-ab8d-f2c67912da66 req-06ee71df-65c1-40c8-b460-68d2213a9315 service nova] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Refreshing instance network info cache due to event network-changed-6ca336c0-76d8-4913-91e7-f7c65a3aa5c1. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 884.728335] env[62569]: DEBUG oslo_concurrency.lockutils [req-7f96f7f4-043a-45f1-ab8d-f2c67912da66 req-06ee71df-65c1-40c8-b460-68d2213a9315 service nova] Acquiring lock "refresh_cache-9c1c0379-3968-464b-b587-8fac704404d5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.728484] env[62569]: DEBUG oslo_concurrency.lockutils [req-7f96f7f4-043a-45f1-ab8d-f2c67912da66 req-06ee71df-65c1-40c8-b460-68d2213a9315 service nova] Acquired lock "refresh_cache-9c1c0379-3968-464b-b587-8fac704404d5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.728651] env[62569]: DEBUG nova.network.neutron [req-7f96f7f4-043a-45f1-ab8d-f2c67912da66 req-06ee71df-65c1-40c8-b460-68d2213a9315 service nova] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Refreshing network info cache for port 6ca336c0-76d8-4913-91e7-f7c65a3aa5c1 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 884.762033] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 884.762033] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 884.762033] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Deleting the datastore file [datastore2] 298efba9-226d-4105-8b32-76b3dddddb0b {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 884.762033] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-18667aae-ad7f-4754-8d04-4761894a2fc9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.768281] env[62569]: DEBUG oslo_vmware.api [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Waiting for the task: (returnval){ [ 884.768281] env[62569]: value = "task-1250286" [ 884.768281] env[62569]: _type = "Task" [ 884.768281] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.777064] env[62569]: DEBUG oslo_vmware.api [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Task: {'id': task-1250286, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.787377] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250282, 'name': CreateVM_Task, 'duration_secs': 0.354652} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.787798] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 884.791061] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5909e6a6-6c00-4956-a4d0-c6430d91b165" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.791061] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5909e6a6-6c00-4956-a4d0-c6430d91b165" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.791061] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5909e6a6-6c00-4956-a4d0-c6430d91b165" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 884.791061] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7ab1ad5-f8de-4353-8a5b-1ca0e8649c73 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.796197] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 884.796197] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]525c8142-7978-50f1-1767-96b4481c8099" [ 884.796197] env[62569]: _type = "Task" [ 884.796197] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.802725] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]525c8142-7978-50f1-1767-96b4481c8099, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.838884] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9781eb1c-e7a2-4f4c-aa82-aeb8a234f46d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "108d5bde-e463-4a69-816b-bdd3c03e13cd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.947718] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250280, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.021512] env[62569]: DEBUG oslo_vmware.api [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250283, 'name': ReconfigVM_Task, 'duration_secs': 0.366206} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.021826] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Updating instance '31ce29fa-4f60-4404-b830-21ad196f78b5' progress to 33 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 885.025689] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.436s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.026235] env[62569]: DEBUG nova.compute.manager [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 885.029031] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.600s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.030264] env[62569]: INFO nova.compute.claims [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 885.060298] env[62569]: DEBUG oslo_vmware.api [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250284, 'name': CloneVM_Task} progress is 94%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.193508] env[62569]: DEBUG oslo_vmware.api [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250270, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.278260] env[62569]: DEBUG oslo_vmware.api [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Task: {'id': task-1250286, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.249362} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.278542] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 885.278808] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 885.279324] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 885.279324] env[62569]: INFO nova.compute.manager [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Took 1.13 seconds to destroy the instance on the hypervisor. [ 885.279656] env[62569]: DEBUG oslo.service.loopingcall [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 885.279825] env[62569]: DEBUG nova.compute.manager [-] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 885.279919] env[62569]: DEBUG nova.network.neutron [-] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 885.303596] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5909e6a6-6c00-4956-a4d0-c6430d91b165" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.303752] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Processing image 5909e6a6-6c00-4956-a4d0-c6430d91b165 {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 885.304010] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5909e6a6-6c00-4956-a4d0-c6430d91b165/5909e6a6-6c00-4956-a4d0-c6430d91b165.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.304170] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5909e6a6-6c00-4956-a4d0-c6430d91b165/5909e6a6-6c00-4956-a4d0-c6430d91b165.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.304383] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 885.304650] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fa7522a0-5206-4e68-a5ec-23f3ce89bb18 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.312324] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 885.312614] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 885.313231] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7c3a725-4528-470b-9b89-33d8442a8fcc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.318239] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 885.318239] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52e37781-4c66-e335-f8ad-67eb882250a8" [ 885.318239] env[62569]: _type = "Task" [ 885.318239] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.331726] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Preparing fetch location {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 885.332022] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Fetch image to [datastore1] OSTACK_IMG_bead562b-e9db-4a1b-b619-67588ba07eae/OSTACK_IMG_bead562b-e9db-4a1b-b619-67588ba07eae.vmdk {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 885.333021] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Downloading stream optimized image 5909e6a6-6c00-4956-a4d0-c6430d91b165 to [datastore1] OSTACK_IMG_bead562b-e9db-4a1b-b619-67588ba07eae/OSTACK_IMG_bead562b-e9db-4a1b-b619-67588ba07eae.vmdk on the data store datastore1 as vApp {{(pid=62569) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 885.333243] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Downloading image file data 5909e6a6-6c00-4956-a4d0-c6430d91b165 to the ESX as VM named 'OSTACK_IMG_bead562b-e9db-4a1b-b619-67588ba07eae' {{(pid=62569) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 885.418325] env[62569]: DEBUG oslo_vmware.rw_handles [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 885.418325] env[62569]: value = "resgroup-9" [ 885.418325] env[62569]: _type = "ResourcePool" [ 885.418325] env[62569]: }. {{(pid=62569) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 885.418783] env[62569]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-fb037680-4ac0-4be8-ab29-05443483cbff {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.452962] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250280, 'name': ReconfigVM_Task, 'duration_secs': 1.61043} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.454144] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 337d88a2-30b9-4846-929e-042bd7a64a65/337d88a2-30b9-4846-929e-042bd7a64a65.vmdk or device None with type streamOptimized {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 885.455068] env[62569]: DEBUG oslo_vmware.rw_handles [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lease: (returnval){ [ 885.455068] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5223dda5-4e48-3cb0-0de5-0829601983e1" [ 885.455068] env[62569]: _type = "HttpNfcLease" [ 885.455068] env[62569]: } obtained for vApp import into resource pool (val){ [ 885.455068] env[62569]: value = "resgroup-9" [ 885.455068] env[62569]: _type = "ResourcePool" [ 885.455068] env[62569]: }. {{(pid=62569) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 885.455616] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the lease: (returnval){ [ 885.455616] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5223dda5-4e48-3cb0-0de5-0829601983e1" [ 885.455616] env[62569]: _type = "HttpNfcLease" [ 885.455616] env[62569]: } to be ready. {{(pid=62569) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 885.455616] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-617513b4-efb8-4c43-a866-86962bf3cd53 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.463972] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 885.463972] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5223dda5-4e48-3cb0-0de5-0829601983e1" [ 885.463972] env[62569]: _type = "HttpNfcLease" [ 885.463972] env[62569]: } is initializing. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 885.465427] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 885.465427] env[62569]: value = "task-1250288" [ 885.465427] env[62569]: _type = "Task" [ 885.465427] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.475847] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250288, 'name': Rename_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.534263] env[62569]: DEBUG nova.virt.hardware [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 885.534263] env[62569]: DEBUG nova.virt.hardware [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 885.534263] env[62569]: DEBUG nova.virt.hardware [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 885.534263] env[62569]: DEBUG nova.virt.hardware [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 885.534263] env[62569]: DEBUG nova.virt.hardware [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 885.534263] env[62569]: DEBUG nova.virt.hardware [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 885.534263] env[62569]: DEBUG nova.virt.hardware [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 885.534515] env[62569]: DEBUG nova.virt.hardware [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 885.534734] env[62569]: DEBUG nova.virt.hardware [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 885.534988] env[62569]: DEBUG nova.virt.hardware [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 885.535283] env[62569]: DEBUG nova.virt.hardware [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 885.542559] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Reconfiguring VM instance instance-0000004f to detach disk 2000 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 885.544095] env[62569]: DEBUG nova.compute.utils [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 885.552216] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0cd5766b-04ab-4228-8f5d-8914bdacc127 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.572846] env[62569]: DEBUG nova.compute.manager [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 885.573104] env[62569]: DEBUG nova.network.neutron [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 885.583334] env[62569]: DEBUG nova.network.neutron [req-7f96f7f4-043a-45f1-ab8d-f2c67912da66 req-06ee71df-65c1-40c8-b460-68d2213a9315 service nova] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Updated VIF entry in instance network info cache for port 6ca336c0-76d8-4913-91e7-f7c65a3aa5c1. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 885.583796] env[62569]: DEBUG nova.network.neutron [req-7f96f7f4-043a-45f1-ab8d-f2c67912da66 req-06ee71df-65c1-40c8-b460-68d2213a9315 service nova] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Updating instance_info_cache with network_info: [{"id": "6ca336c0-76d8-4913-91e7-f7c65a3aa5c1", "address": "fa:16:3e:b8:41:ba", "network": {"id": "8ea5e138-813b-4c5c-88a1-b1ea79807f10", "bridge": "br-int", "label": "tempest-ImagesTestJSON-524163445-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b20340a1ce0447cae1bfd8b7c28928f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ca336c0-76", "ovs_interfaceid": "6ca336c0-76d8-4913-91e7-f7c65a3aa5c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.591115] env[62569]: DEBUG oslo_vmware.api [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 885.591115] env[62569]: value = "task-1250289" [ 885.591115] env[62569]: _type = "Task" [ 885.591115] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.591441] env[62569]: DEBUG oslo_vmware.api [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250284, 'name': CloneVM_Task} progress is 95%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.605076] env[62569]: DEBUG oslo_vmware.api [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250289, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.696594] env[62569]: DEBUG oslo_vmware.api [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250270, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.887087] env[62569]: DEBUG nova.policy [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6dd052494f244fd78050c5d26f497720', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2c6a4814c234d53bbd4844cd482fbae', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 885.947026] env[62569]: DEBUG nova.compute.manager [req-b383a59f-c512-4ce9-ab2b-2b929cd7a6ac req-9bb8c6b8-49f7-46e3-970c-b6624ed538bb service nova] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Received event network-vif-plugged-057050a7-da70-4b2f-bd56-b5de3a3dbb31 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 885.947026] env[62569]: DEBUG oslo_concurrency.lockutils [req-b383a59f-c512-4ce9-ab2b-2b929cd7a6ac req-9bb8c6b8-49f7-46e3-970c-b6624ed538bb service nova] Acquiring lock "a6ee3547-3a3d-4480-b451-d1fddd829b39-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.947026] env[62569]: DEBUG oslo_concurrency.lockutils [req-b383a59f-c512-4ce9-ab2b-2b929cd7a6ac req-9bb8c6b8-49f7-46e3-970c-b6624ed538bb service nova] Lock "a6ee3547-3a3d-4480-b451-d1fddd829b39-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.947286] env[62569]: DEBUG oslo_concurrency.lockutils [req-b383a59f-c512-4ce9-ab2b-2b929cd7a6ac req-9bb8c6b8-49f7-46e3-970c-b6624ed538bb service nova] Lock "a6ee3547-3a3d-4480-b451-d1fddd829b39-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.947286] env[62569]: DEBUG nova.compute.manager [req-b383a59f-c512-4ce9-ab2b-2b929cd7a6ac req-9bb8c6b8-49f7-46e3-970c-b6624ed538bb service nova] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] No waiting events found dispatching network-vif-plugged-057050a7-da70-4b2f-bd56-b5de3a3dbb31 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 885.947561] env[62569]: WARNING nova.compute.manager [req-b383a59f-c512-4ce9-ab2b-2b929cd7a6ac req-9bb8c6b8-49f7-46e3-970c-b6624ed538bb service nova] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Received unexpected event network-vif-plugged-057050a7-da70-4b2f-bd56-b5de3a3dbb31 for instance with vm_state building and task_state spawning. [ 885.967337] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 885.967337] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5223dda5-4e48-3cb0-0de5-0829601983e1" [ 885.967337] env[62569]: _type = "HttpNfcLease" [ 885.967337] env[62569]: } is initializing. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 885.977836] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5b6963b5-bcc3-4f7e-aa82-4e266677d1b7 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "108d5bde-e463-4a69-816b-bdd3c03e13cd" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.978103] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5b6963b5-bcc3-4f7e-aa82-4e266677d1b7 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "108d5bde-e463-4a69-816b-bdd3c03e13cd" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.983634] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250288, 'name': Rename_Task, 'duration_secs': 0.157929} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.986331] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 885.986331] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8e18dfe3-ea19-4c78-9a3e-c1866ddd62f1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.992726] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 885.992726] env[62569]: value = "task-1250290" [ 885.992726] env[62569]: _type = "Task" [ 885.992726] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.004181] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250290, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.075750] env[62569]: DEBUG nova.compute.manager [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 886.091389] env[62569]: DEBUG oslo_concurrency.lockutils [req-7f96f7f4-043a-45f1-ab8d-f2c67912da66 req-06ee71df-65c1-40c8-b460-68d2213a9315 service nova] Releasing lock "refresh_cache-9c1c0379-3968-464b-b587-8fac704404d5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.091799] env[62569]: DEBUG oslo_vmware.api [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250284, 'name': CloneVM_Task, 'duration_secs': 1.223106} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.092133] env[62569]: INFO nova.virt.vmwareapi.vmops [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Created linked-clone VM from snapshot [ 886.095722] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c2fd520-905b-4c91-8a9d-9a8af119d52b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.107751] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Uploading image a618f103-681b-4bbf-a4d2-f503940d1cf8 {{(pid=62569) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 886.109726] env[62569]: DEBUG oslo_vmware.api [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250289, 'name': ReconfigVM_Task, 'duration_secs': 0.174334} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.110214] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Reconfigured VM instance instance-0000004f to detach disk 2000 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 886.111090] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bae83ce-0dfa-4cc8-a940-6fcb30074040 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.134329] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 31ce29fa-4f60-4404-b830-21ad196f78b5/31ce29fa-4f60-4404-b830-21ad196f78b5.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 886.139454] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8362b088-c3c7-48ed-97fb-a55c588260c0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.159132] env[62569]: DEBUG oslo_vmware.rw_handles [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 886.159132] env[62569]: value = "vm-269455" [ 886.159132] env[62569]: _type = "VirtualMachine" [ 886.159132] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 886.160297] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-48406f40-87fe-4dfc-a172-91b1bd6facaf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.163563] env[62569]: DEBUG oslo_vmware.api [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 886.163563] env[62569]: value = "task-1250291" [ 886.163563] env[62569]: _type = "Task" [ 886.163563] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.166655] env[62569]: DEBUG nova.network.neutron [-] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.173399] env[62569]: DEBUG oslo_vmware.rw_handles [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lease: (returnval){ [ 886.173399] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528d6e16-6d77-06d5-0e4b-fbdf3be3f5f1" [ 886.173399] env[62569]: _type = "HttpNfcLease" [ 886.173399] env[62569]: } obtained for exporting VM: (result){ [ 886.173399] env[62569]: value = "vm-269455" [ 886.173399] env[62569]: _type = "VirtualMachine" [ 886.173399] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 886.173704] env[62569]: DEBUG oslo_vmware.api [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the lease: (returnval){ [ 886.173704] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528d6e16-6d77-06d5-0e4b-fbdf3be3f5f1" [ 886.173704] env[62569]: _type = "HttpNfcLease" [ 886.173704] env[62569]: } to be ready. {{(pid=62569) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 886.182291] env[62569]: DEBUG oslo_vmware.api [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250291, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.191439] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 886.191439] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528d6e16-6d77-06d5-0e4b-fbdf3be3f5f1" [ 886.191439] env[62569]: _type = "HttpNfcLease" [ 886.191439] env[62569]: } is ready. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 886.192133] env[62569]: DEBUG oslo_vmware.rw_handles [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 886.192133] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528d6e16-6d77-06d5-0e4b-fbdf3be3f5f1" [ 886.192133] env[62569]: _type = "HttpNfcLease" [ 886.192133] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 886.192901] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7b01e4a-9250-4a80-8291-96d684c5a6a2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.199181] env[62569]: DEBUG oslo_vmware.api [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250270, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.208235] env[62569]: DEBUG oslo_vmware.rw_handles [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523457f5-4c39-ea8d-cb0a-93de5f052b50/disk-0.vmdk from lease info. {{(pid=62569) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 886.208449] env[62569]: DEBUG oslo_vmware.rw_handles [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523457f5-4c39-ea8d-cb0a-93de5f052b50/disk-0.vmdk for reading. {{(pid=62569) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 886.303077] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-614db23d-824f-408f-90db-f5a8a17b02c7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.378810] env[62569]: DEBUG nova.network.neutron [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Successfully created port: 37c7a902-07de-45dc-b683-311b84a5edc5 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 886.465271] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d10799e8-dc20-416e-a15c-6f94bbfa3758 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.474248] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 886.474248] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5223dda5-4e48-3cb0-0de5-0829601983e1" [ 886.474248] env[62569]: _type = "HttpNfcLease" [ 886.474248] env[62569]: } is ready. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 886.475273] env[62569]: DEBUG nova.network.neutron [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Successfully updated port: 057050a7-da70-4b2f-bd56-b5de3a3dbb31 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 886.479239] env[62569]: DEBUG oslo_vmware.rw_handles [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 886.479239] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5223dda5-4e48-3cb0-0de5-0829601983e1" [ 886.479239] env[62569]: _type = "HttpNfcLease" [ 886.479239] env[62569]: }. {{(pid=62569) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 886.480563] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea06aba0-0992-425f-8de9-eafda20bb814 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.484723] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82f56ee8-5f09-44c2-99e9-1226f4d0bb18 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.492578] env[62569]: DEBUG nova.compute.utils [None req-5b6963b5-bcc3-4f7e-aa82-4e266677d1b7 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 886.536432] env[62569]: DEBUG oslo_vmware.rw_handles [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52077886-9aab-95f8-92e2-c1599928285a/disk-0.vmdk from lease info. {{(pid=62569) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 886.536831] env[62569]: DEBUG oslo_vmware.rw_handles [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52077886-9aab-95f8-92e2-c1599928285a/disk-0.vmdk. {{(pid=62569) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 886.539247] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b3e566b-9045-4569-931c-024cbca7d4e6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.547269] env[62569]: DEBUG oslo_vmware.api [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250290, 'name': PowerOnVM_Task, 'duration_secs': 0.540387} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.604277] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 886.619032] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caab06f7-273a-4408-a161-c5c42e4e0d44 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.623972] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8fb3bd94-00ce-45c1-998c-46d92a556d12 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.636423] env[62569]: DEBUG nova.compute.provider_tree [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.673499] env[62569]: INFO nova.compute.manager [-] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Took 1.39 seconds to deallocate network for instance. [ 886.680512] env[62569]: DEBUG oslo_vmware.api [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250291, 'name': ReconfigVM_Task, 'duration_secs': 0.306741} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.683662] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 31ce29fa-4f60-4404-b830-21ad196f78b5/31ce29fa-4f60-4404-b830-21ad196f78b5.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 886.683854] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Updating instance '31ce29fa-4f60-4404-b830-21ad196f78b5' progress to 50 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 886.696608] env[62569]: DEBUG oslo_vmware.api [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250270, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.720716] env[62569]: DEBUG nova.compute.manager [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 886.721686] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77b919e3-fb14-476b-b6bf-0c1bba8163af {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.982820] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Acquiring lock "refresh_cache-a6ee3547-3a3d-4480-b451-d1fddd829b39" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.982820] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Acquired lock "refresh_cache-a6ee3547-3a3d-4480-b451-d1fddd829b39" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.982820] env[62569]: DEBUG nova.network.neutron [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 886.996542] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5b6963b5-bcc3-4f7e-aa82-4e266677d1b7 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "108d5bde-e463-4a69-816b-bdd3c03e13cd" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.018s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.064186] env[62569]: DEBUG nova.compute.manager [req-cb6262b0-e2fc-4339-b8b4-3dfb4e674080 req-79c63d95-2c5c-43e2-aece-5912dea8f1b7 service nova] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Received event network-changed-057050a7-da70-4b2f-bd56-b5de3a3dbb31 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 887.064435] env[62569]: DEBUG nova.compute.manager [req-cb6262b0-e2fc-4339-b8b4-3dfb4e674080 req-79c63d95-2c5c-43e2-aece-5912dea8f1b7 service nova] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Refreshing instance network info cache due to event network-changed-057050a7-da70-4b2f-bd56-b5de3a3dbb31. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 887.064605] env[62569]: DEBUG oslo_concurrency.lockutils [req-cb6262b0-e2fc-4339-b8b4-3dfb4e674080 req-79c63d95-2c5c-43e2-aece-5912dea8f1b7 service nova] Acquiring lock "refresh_cache-a6ee3547-3a3d-4480-b451-d1fddd829b39" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.087223] env[62569]: DEBUG nova.compute.manager [req-94271c77-9623-4d67-8873-ede07adceb97 req-b46e6f34-23f9-4726-a135-844e952326c6 service nova] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Received event network-vif-deleted-f33064d8-b284-4ed3-9a2b-01f27371be83 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 887.119162] env[62569]: DEBUG nova.compute.manager [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 887.141671] env[62569]: DEBUG nova.scheduler.client.report [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 887.147851] env[62569]: DEBUG nova.virt.hardware [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 887.149114] env[62569]: DEBUG nova.virt.hardware [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 887.149114] env[62569]: DEBUG nova.virt.hardware [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 887.149114] env[62569]: DEBUG nova.virt.hardware [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 887.149114] env[62569]: DEBUG nova.virt.hardware [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 887.149114] env[62569]: DEBUG nova.virt.hardware [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 887.149412] env[62569]: DEBUG nova.virt.hardware [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 887.149535] env[62569]: DEBUG nova.virt.hardware [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 887.149848] env[62569]: DEBUG nova.virt.hardware [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 887.150108] env[62569]: DEBUG nova.virt.hardware [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 887.150346] env[62569]: DEBUG nova.virt.hardware [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 887.154871] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-748f9acf-de11-4d19-8a64-4900b8890695 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.169102] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b237044-88b2-453e-a34e-48b3d42b2261 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.191100] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.196521] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4694ec41-63cb-40d9-be81-52f67e0e5aea {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.209255] env[62569]: DEBUG oslo_vmware.api [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250270, 'name': ReconfigVM_Task, 'duration_secs': 5.950372} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.232329] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.232717] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Reconfigured VM to detach interface {{(pid=62569) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 887.246179] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8cd4180-1e39-4619-88c3-3e7822ee7c3a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.252329] env[62569]: DEBUG oslo_concurrency.lockutils [None req-057bfd07-d3e6-4113-9573-fa4ad159f341 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "337d88a2-30b9-4846-929e-042bd7a64a65" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 24.559s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.279039] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Updating instance '31ce29fa-4f60-4404-b830-21ad196f78b5' progress to 67 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 887.538075] env[62569]: DEBUG nova.network.neutron [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 887.579020] env[62569]: DEBUG oslo_vmware.rw_handles [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Completed reading data from the image iterator. {{(pid=62569) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 887.579020] env[62569]: DEBUG oslo_vmware.rw_handles [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52077886-9aab-95f8-92e2-c1599928285a/disk-0.vmdk. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 887.579020] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-490e568d-519c-4b74-97df-8b746326f7d6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.588799] env[62569]: DEBUG oslo_vmware.rw_handles [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52077886-9aab-95f8-92e2-c1599928285a/disk-0.vmdk is in state: ready. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 887.589196] env[62569]: DEBUG oslo_vmware.rw_handles [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52077886-9aab-95f8-92e2-c1599928285a/disk-0.vmdk. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 887.589566] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-8d286412-7bcb-4d67-a594-52eb6f6c7f89 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.659746] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.631s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.660322] env[62569]: DEBUG nova.compute.manager [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 887.663748] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.598s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.664076] env[62569]: DEBUG nova.objects.instance [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Lazy-loading 'resources' on Instance uuid 691d8fbe-b9ee-454f-bd7b-14520e53ed26 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.727091] env[62569]: DEBUG nova.network.neutron [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Updating instance_info_cache with network_info: [{"id": "057050a7-da70-4b2f-bd56-b5de3a3dbb31", "address": "fa:16:3e:67:62:98", "network": {"id": "2e40c5b0-afc9-4e2a-babe-cdad255fb024", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-865392953-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "335bc4885a9e4675a3357199a612faed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea45c024-d603-4bac-9c1b-f302437ea4fe", "external-id": "nsx-vlan-transportzone-946", "segmentation_id": 946, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap057050a7-da", "ovs_interfaceid": "057050a7-da70-4b2f-bd56-b5de3a3dbb31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.826468] env[62569]: DEBUG nova.network.neutron [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Port 8e55dca6-f777-4333-9b49-feb5527d9a5e binding to destination host cpu-1 is already ACTIVE {{(pid=62569) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 887.979324] env[62569]: DEBUG oslo_vmware.rw_handles [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52077886-9aab-95f8-92e2-c1599928285a/disk-0.vmdk. {{(pid=62569) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 887.979605] env[62569]: INFO nova.virt.vmwareapi.images [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Downloaded image file data 5909e6a6-6c00-4956-a4d0-c6430d91b165 [ 887.987023] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b246c3d-dc75-4654-9340-777da4a7d9eb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.006019] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e8ef703c-3236-47ec-8b9d-976bba852184 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.044867] env[62569]: INFO nova.virt.vmwareapi.images [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] The imported VM was unregistered [ 888.047853] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Caching image {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 888.048291] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Creating directory with path [datastore1] devstack-image-cache_base/5909e6a6-6c00-4956-a4d0-c6430d91b165 {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 888.048954] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2353079c-89dd-4aed-bd06-f17a6a3f1619 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.063782] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Created directory with path [datastore1] devstack-image-cache_base/5909e6a6-6c00-4956-a4d0-c6430d91b165 {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 888.064015] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_bead562b-e9db-4a1b-b619-67588ba07eae/OSTACK_IMG_bead562b-e9db-4a1b-b619-67588ba07eae.vmdk to [datastore1] devstack-image-cache_base/5909e6a6-6c00-4956-a4d0-c6430d91b165/5909e6a6-6c00-4956-a4d0-c6430d91b165.vmdk. {{(pid=62569) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 888.064293] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-5ae01ed4-3513-49a1-b9b7-8f36bda5d048 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.068498] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5b6963b5-bcc3-4f7e-aa82-4e266677d1b7 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "108d5bde-e463-4a69-816b-bdd3c03e13cd" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.068784] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5b6963b5-bcc3-4f7e-aa82-4e266677d1b7 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "108d5bde-e463-4a69-816b-bdd3c03e13cd" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.069038] env[62569]: INFO nova.compute.manager [None req-5b6963b5-bcc3-4f7e-aa82-4e266677d1b7 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Attaching volume d20ca906-3332-4f05-bd83-1a0427d2b692 to /dev/sdb [ 888.079083] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 888.079083] env[62569]: value = "task-1250294" [ 888.079083] env[62569]: _type = "Task" [ 888.079083] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.088217] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250294, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.105378] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8149bb36-3756-4a83-b5a4-b1b97901cc81 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.116751] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1a687b5-58da-409a-b587-36113542410d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.131864] env[62569]: DEBUG nova.virt.block_device [None req-5b6963b5-bcc3-4f7e-aa82-4e266677d1b7 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Updating existing volume attachment record: b03d88cb-e6c8-4040-b428-75d9777f3f88 {{(pid=62569) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 888.171535] env[62569]: DEBUG nova.compute.utils [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 888.177227] env[62569]: DEBUG nova.compute.manager [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 888.177227] env[62569]: DEBUG nova.network.neutron [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 888.231236] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Releasing lock "refresh_cache-a6ee3547-3a3d-4480-b451-d1fddd829b39" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.232369] env[62569]: DEBUG nova.compute.manager [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Instance network_info: |[{"id": "057050a7-da70-4b2f-bd56-b5de3a3dbb31", "address": "fa:16:3e:67:62:98", "network": {"id": "2e40c5b0-afc9-4e2a-babe-cdad255fb024", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-865392953-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "335bc4885a9e4675a3357199a612faed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea45c024-d603-4bac-9c1b-f302437ea4fe", "external-id": "nsx-vlan-transportzone-946", "segmentation_id": 946, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap057050a7-da", "ovs_interfaceid": "057050a7-da70-4b2f-bd56-b5de3a3dbb31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 888.232369] env[62569]: DEBUG oslo_concurrency.lockutils [req-cb6262b0-e2fc-4339-b8b4-3dfb4e674080 req-79c63d95-2c5c-43e2-aece-5912dea8f1b7 service nova] Acquired lock "refresh_cache-a6ee3547-3a3d-4480-b451-d1fddd829b39" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.232369] env[62569]: DEBUG nova.network.neutron [req-cb6262b0-e2fc-4339-b8b4-3dfb4e674080 req-79c63d95-2c5c-43e2-aece-5912dea8f1b7 service nova] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Refreshing network info cache for port 057050a7-da70-4b2f-bd56-b5de3a3dbb31 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 888.237147] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:67:62:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ea45c024-d603-4bac-9c1b-f302437ea4fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '057050a7-da70-4b2f-bd56-b5de3a3dbb31', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 888.247549] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Creating folder: Project (335bc4885a9e4675a3357199a612faed). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 888.252255] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0467f273-1ec3-41c9-ac5c-8480f0cdd137 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.270119] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Created folder: Project (335bc4885a9e4675a3357199a612faed) in parent group-v269330. [ 888.270119] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Creating folder: Instances. Parent ref: group-v269457. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 888.270119] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-53673f63-b394-4253-b4bc-30ac142ddec3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.281852] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Created folder: Instances in parent group-v269457. [ 888.282319] env[62569]: DEBUG oslo.service.loopingcall [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 888.284911] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 888.285461] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9729ccbe-6e0f-4664-b49e-97ef64013574 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.301768] env[62569]: DEBUG nova.policy [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '50780a0da3b14122aa6287c1a482f671', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7f7fb3e698324f8e851aefbee74ac669', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 888.307033] env[62569]: DEBUG nova.network.neutron [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Successfully updated port: 37c7a902-07de-45dc-b683-311b84a5edc5 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 888.317092] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 888.317092] env[62569]: value = "task-1250298" [ 888.317092] env[62569]: _type = "Task" [ 888.317092] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.327429] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250298, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.532764] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0bd08b6-1255-4b05-a97a-8fbeaf8f5ff4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.545281] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-842faf9b-75fc-4fb1-af5a-b653826faf44 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.585300] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a94b13ba-d072-4383-91bd-98a2cb952256 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.597900] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-627cc544-3235-418b-9134-4b8d8fc4f398 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.602412] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250294, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.618629] env[62569]: DEBUG nova.compute.provider_tree [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.649457] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.649722] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.650031] env[62569]: DEBUG nova.network.neutron [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 888.675703] env[62569]: DEBUG nova.compute.manager [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 888.791244] env[62569]: DEBUG nova.network.neutron [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Successfully created port: d64908ff-b5fa-42fe-8a49-1f39a27ca4b7 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 888.810847] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "refresh_cache-534c9ece-c8df-4528-83d8-c6c0ef8c4793" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.811216] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired lock "refresh_cache-534c9ece-c8df-4528-83d8-c6c0ef8c4793" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.811542] env[62569]: DEBUG nova.network.neutron [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 888.833416] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250298, 'name': CreateVM_Task, 'duration_secs': 0.495012} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.834466] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 888.835267] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.835500] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.835869] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 888.844824] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1f5a649-dd68-44f2-84a2-3a9205265f6a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.854415] env[62569]: DEBUG oslo_concurrency.lockutils [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "31ce29fa-4f60-4404-b830-21ad196f78b5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.854713] env[62569]: DEBUG oslo_concurrency.lockutils [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "31ce29fa-4f60-4404-b830-21ad196f78b5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.854935] env[62569]: DEBUG oslo_concurrency.lockutils [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "31ce29fa-4f60-4404-b830-21ad196f78b5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.863069] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Waiting for the task: (returnval){ [ 888.863069] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529f32ff-ce24-de71-c01d-4fecac0641a3" [ 888.863069] env[62569]: _type = "Task" [ 888.863069] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.877096] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529f32ff-ce24-de71-c01d-4fecac0641a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.096285] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250294, 'name': MoveVirtualDisk_Task} progress is 35%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.129359] env[62569]: DEBUG nova.scheduler.client.report [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 889.134868] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "807d3025-d6a7-4778-a829-a61e2c7495c9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.135339] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "807d3025-d6a7-4778-a829-a61e2c7495c9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.135678] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "807d3025-d6a7-4778-a829-a61e2c7495c9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.135977] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "807d3025-d6a7-4778-a829-a61e2c7495c9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.136311] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "807d3025-d6a7-4778-a829-a61e2c7495c9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.139240] env[62569]: INFO nova.compute.manager [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Terminating instance [ 889.244235] env[62569]: DEBUG nova.network.neutron [req-cb6262b0-e2fc-4339-b8b4-3dfb4e674080 req-79c63d95-2c5c-43e2-aece-5912dea8f1b7 service nova] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Updated VIF entry in instance network info cache for port 057050a7-da70-4b2f-bd56-b5de3a3dbb31. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 889.244803] env[62569]: DEBUG nova.network.neutron [req-cb6262b0-e2fc-4339-b8b4-3dfb4e674080 req-79c63d95-2c5c-43e2-aece-5912dea8f1b7 service nova] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Updating instance_info_cache with network_info: [{"id": "057050a7-da70-4b2f-bd56-b5de3a3dbb31", "address": "fa:16:3e:67:62:98", "network": {"id": "2e40c5b0-afc9-4e2a-babe-cdad255fb024", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-865392953-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "335bc4885a9e4675a3357199a612faed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea45c024-d603-4bac-9c1b-f302437ea4fe", "external-id": "nsx-vlan-transportzone-946", "segmentation_id": 946, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap057050a7-da", "ovs_interfaceid": "057050a7-da70-4b2f-bd56-b5de3a3dbb31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.381042] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529f32ff-ce24-de71-c01d-4fecac0641a3, 'name': SearchDatastore_Task, 'duration_secs': 0.094321} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.381747] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.382294] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 889.382744] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.383039] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.383353] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 889.383763] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-49eda48f-ad08-4054-b135-d47fc12ed089 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.404096] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 889.404657] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 889.405554] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71796f05-6b38-43fd-abf2-1e1683310a2e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.409354] env[62569]: DEBUG nova.network.neutron [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 889.419495] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Waiting for the task: (returnval){ [ 889.419495] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f5533e-5081-125c-b013-f3f1c705c9fe" [ 889.419495] env[62569]: _type = "Task" [ 889.419495] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.432552] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f5533e-5081-125c-b013-f3f1c705c9fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.489580] env[62569]: DEBUG nova.compute.manager [req-7affada8-b78d-45ff-be6a-e5d7fca572e2 req-27a73008-78ff-48fa-b162-fe3fca2d1efc service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Received event network-vif-deleted-5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 889.490206] env[62569]: INFO nova.compute.manager [req-7affada8-b78d-45ff-be6a-e5d7fca572e2 req-27a73008-78ff-48fa-b162-fe3fca2d1efc service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Neutron deleted interface 5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18; detaching it from the instance and deleting it from the info cache [ 889.491247] env[62569]: DEBUG nova.network.neutron [req-7affada8-b78d-45ff-be6a-e5d7fca572e2 req-27a73008-78ff-48fa-b162-fe3fca2d1efc service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Updating instance_info_cache with network_info: [{"id": "9d855102-33ab-4574-b791-dfa6fb5b0f09", "address": "fa:16:3e:b3:4b:f5", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d855102-33", "ovs_interfaceid": "9d855102-33ab-4574-b791-dfa6fb5b0f09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "62c1052e-dc0a-4fb6-a399-3e238bf93789", "address": "fa:16:3e:9d:69:9f", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62c1052e-dc", "ovs_interfaceid": "62c1052e-dc0a-4fb6-a399-3e238bf93789", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.513499] env[62569]: DEBUG nova.compute.manager [req-2ba108f8-5e2b-4a70-8613-12bbca19a584 req-0973b397-db6e-4d35-87d6-7273a151ce44 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Received event network-vif-deleted-62c1052e-dc0a-4fb6-a399-3e238bf93789 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 889.513894] env[62569]: INFO nova.compute.manager [req-2ba108f8-5e2b-4a70-8613-12bbca19a584 req-0973b397-db6e-4d35-87d6-7273a151ce44 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Neutron deleted interface 62c1052e-dc0a-4fb6-a399-3e238bf93789; detaching it from the instance and deleting it from the info cache [ 889.514450] env[62569]: DEBUG nova.network.neutron [req-2ba108f8-5e2b-4a70-8613-12bbca19a584 req-0973b397-db6e-4d35-87d6-7273a151ce44 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Updating instance_info_cache with network_info: [{"id": "9d855102-33ab-4574-b791-dfa6fb5b0f09", "address": "fa:16:3e:b3:4b:f5", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d855102-33", "ovs_interfaceid": "9d855102-33ab-4574-b791-dfa6fb5b0f09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18", "address": "fa:16:3e:df:90:ac", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f2ad2ec-3e", "ovs_interfaceid": "5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.600531] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250294, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.641624] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.976s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.642566] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.452s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.643327] env[62569]: DEBUG nova.objects.instance [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Lazy-loading 'resources' on Instance uuid 298efba9-226d-4105-8b32-76b3dddddb0b {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 889.645130] env[62569]: DEBUG nova.compute.manager [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 889.645529] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 889.650017] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d26a37ad-3ac5-4e96-b509-a9e432a23fb8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.663955] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 889.666316] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-678df121-cfd4-4800-acdb-725350135a8d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.668954] env[62569]: INFO nova.scheduler.client.report [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Deleted allocations for instance 691d8fbe-b9ee-454f-bd7b-14520e53ed26 [ 889.682990] env[62569]: DEBUG oslo_vmware.api [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 889.682990] env[62569]: value = "task-1250301" [ 889.682990] env[62569]: _type = "Task" [ 889.682990] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.686865] env[62569]: DEBUG nova.compute.manager [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 889.699963] env[62569]: DEBUG oslo_vmware.api [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250301, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.725317] env[62569]: DEBUG nova.virt.hardware [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 889.725866] env[62569]: DEBUG nova.virt.hardware [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 889.726357] env[62569]: DEBUG nova.virt.hardware [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 889.726716] env[62569]: DEBUG nova.virt.hardware [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 889.730023] env[62569]: DEBUG nova.virt.hardware [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 889.730023] env[62569]: DEBUG nova.virt.hardware [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 889.730023] env[62569]: DEBUG nova.virt.hardware [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 889.730023] env[62569]: DEBUG nova.virt.hardware [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 889.730023] env[62569]: DEBUG nova.virt.hardware [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 889.730023] env[62569]: DEBUG nova.virt.hardware [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 889.730023] env[62569]: DEBUG nova.virt.hardware [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 889.730023] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3164e7f5-359b-43ad-a0ac-c7410551f465 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.743592] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a540ca6a-78cb-486d-aae8-0a3b30004aea {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.751599] env[62569]: DEBUG oslo_concurrency.lockutils [req-cb6262b0-e2fc-4339-b8b4-3dfb4e674080 req-79c63d95-2c5c-43e2-aece-5912dea8f1b7 service nova] Releasing lock "refresh_cache-a6ee3547-3a3d-4480-b451-d1fddd829b39" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.870899] env[62569]: DEBUG nova.network.neutron [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Updating instance_info_cache with network_info: [{"id": "37c7a902-07de-45dc-b683-311b84a5edc5", "address": "fa:16:3e:e3:66:93", "network": {"id": "884d3248-07f1-4cf7-8ceb-b94b34622384", "bridge": "br-int", "label": "tempest-ServersTestJSON-1882023778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2c6a4814c234d53bbd4844cd482fbae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37c7a902-07", "ovs_interfaceid": "37c7a902-07de-45dc-b683-311b84a5edc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.904698] env[62569]: INFO nova.network.neutron [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Port 5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 889.904698] env[62569]: INFO nova.network.neutron [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Port 62c1052e-dc0a-4fb6-a399-3e238bf93789 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 889.904698] env[62569]: DEBUG nova.network.neutron [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Updating instance_info_cache with network_info: [{"id": "9d855102-33ab-4574-b791-dfa6fb5b0f09", "address": "fa:16:3e:b3:4b:f5", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d855102-33", "ovs_interfaceid": "9d855102-33ab-4574-b791-dfa6fb5b0f09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.907705] env[62569]: DEBUG oslo_concurrency.lockutils [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "refresh_cache-31ce29fa-4f60-4404-b830-21ad196f78b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.908215] env[62569]: DEBUG oslo_concurrency.lockutils [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired lock "refresh_cache-31ce29fa-4f60-4404-b830-21ad196f78b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.908525] env[62569]: DEBUG nova.network.neutron [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 889.936406] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f5533e-5081-125c-b013-f3f1c705c9fe, 'name': SearchDatastore_Task, 'duration_secs': 0.089035} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.937780] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c64551f-47fd-4c4e-af72-d1fe6a6c8b92 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.949040] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Waiting for the task: (returnval){ [ 889.949040] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52af562f-898a-6cf4-7f8a-12636d26080b" [ 889.949040] env[62569]: _type = "Task" [ 889.949040] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.964839] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52af562f-898a-6cf4-7f8a-12636d26080b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.994773] env[62569]: DEBUG oslo_concurrency.lockutils [req-7affada8-b78d-45ff-be6a-e5d7fca572e2 req-27a73008-78ff-48fa-b162-fe3fca2d1efc service nova] Acquiring lock "807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.020266] env[62569]: DEBUG oslo_concurrency.lockutils [req-2ba108f8-5e2b-4a70-8613-12bbca19a584 req-0973b397-db6e-4d35-87d6-7273a151ce44 service nova] Acquiring lock "807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.096943] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250294, 'name': MoveVirtualDisk_Task} progress is 74%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.184025] env[62569]: DEBUG oslo_concurrency.lockutils [None req-6a269085-820a-4eca-a674-f01fde88e7b0 tempest-ServerTagsTestJSON-756322494 tempest-ServerTagsTestJSON-756322494-project-member] Lock "691d8fbe-b9ee-454f-bd7b-14520e53ed26" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.833s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.205744] env[62569]: DEBUG oslo_vmware.api [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250301, 'name': PowerOffVM_Task, 'duration_secs': 0.273687} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.206299] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 890.206299] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 890.206548] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6dfbfb75-3089-4e1b-95f4-6b71a2c6e938 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.351216] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 890.351216] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 890.351216] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Deleting the datastore file [datastore2] 807d3025-d6a7-4778-a829-a61e2c7495c9 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 890.351624] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-faf9c469-4ab8-4464-8d87-0ee0aed531d5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.363592] env[62569]: DEBUG oslo_vmware.api [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 890.363592] env[62569]: value = "task-1250303" [ 890.363592] env[62569]: _type = "Task" [ 890.363592] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.377198] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Releasing lock "refresh_cache-534c9ece-c8df-4528-83d8-c6c0ef8c4793" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.377575] env[62569]: DEBUG nova.compute.manager [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Instance network_info: |[{"id": "37c7a902-07de-45dc-b683-311b84a5edc5", "address": "fa:16:3e:e3:66:93", "network": {"id": "884d3248-07f1-4cf7-8ceb-b94b34622384", "bridge": "br-int", "label": "tempest-ServersTestJSON-1882023778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2c6a4814c234d53bbd4844cd482fbae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37c7a902-07", "ovs_interfaceid": "37c7a902-07de-45dc-b683-311b84a5edc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 890.378453] env[62569]: DEBUG oslo_vmware.api [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250303, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.378453] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e3:66:93', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e9fa4744-8702-4973-b911-ee18192a3e4b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '37c7a902-07de-45dc-b683-311b84a5edc5', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 890.389035] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Creating folder: Project (c2c6a4814c234d53bbd4844cd482fbae). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 890.391247] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-821a4a6c-655d-456a-ba4a-4ef1f5aacb6c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.407593] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "refresh_cache-807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.410214] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Created folder: Project (c2c6a4814c234d53bbd4844cd482fbae) in parent group-v269330. [ 890.410446] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Creating folder: Instances. Parent ref: group-v269462. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 890.415790] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bc1be99c-fefe-45fa-b856-11ce8db40bbe {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.439072] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Created folder: Instances in parent group-v269462. [ 890.439072] env[62569]: DEBUG oslo.service.loopingcall [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.439072] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 890.439072] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cd56d8c7-23b1-46ff-b2ea-ba717f0b9095 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.474503] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52af562f-898a-6cf4-7f8a-12636d26080b, 'name': SearchDatastore_Task, 'duration_secs': 0.121414} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.476324] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.476666] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] a6ee3547-3a3d-4480-b451-d1fddd829b39/a6ee3547-3a3d-4480-b451-d1fddd829b39.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 890.477149] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 890.477149] env[62569]: value = "task-1250307" [ 890.477149] env[62569]: _type = "Task" [ 890.477149] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.478271] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b41e1719-6007-4feb-9dd5-b7af65ad6f9c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.481306] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77cea944-f870-48b5-a625-e3de6f2d7f6b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.500044] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9d9a8fb-2257-42bb-8908-3288a5700ffc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.504076] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250307, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.506240] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Waiting for the task: (returnval){ [ 890.506240] env[62569]: value = "task-1250308" [ 890.506240] env[62569]: _type = "Task" [ 890.506240] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.544087] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcda89d8-a4df-442b-bf6c-2c5a6ebe5d4e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.546895] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': task-1250308, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.557215] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8443f99e-8cb9-497d-beb9-7daac7e4b88c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.580024] env[62569]: DEBUG nova.compute.provider_tree [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.596084] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250294, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.662955] env[62569]: DEBUG nova.network.neutron [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Successfully updated port: d64908ff-b5fa-42fe-8a49-1f39a27ca4b7 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 890.875406] env[62569]: DEBUG oslo_vmware.api [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250303, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.373909} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.875792] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 890.875984] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 890.876143] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 890.876336] env[62569]: INFO nova.compute.manager [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Took 1.23 seconds to destroy the instance on the hypervisor. [ 890.876590] env[62569]: DEBUG oslo.service.loopingcall [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.876801] env[62569]: DEBUG nova.compute.manager [-] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 890.876899] env[62569]: DEBUG nova.network.neutron [-] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 890.911647] env[62569]: DEBUG nova.network.neutron [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Updating instance_info_cache with network_info: [{"id": "8e55dca6-f777-4333-9b49-feb5527d9a5e", "address": "fa:16:3e:97:39:72", "network": {"id": "c61b7867-7897-43ae-8f5b-19834c4390ea", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-512066189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ba7e96d226942bd99ab29fe703c6fea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e55dca6-f7", "ovs_interfaceid": "8e55dca6-f777-4333-9b49-feb5527d9a5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.924605] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c25d5f8e-7756-43a5-9baa-e5ef90a636ab tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "interface-807d3025-d6a7-4778-a829-a61e2c7495c9-5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.366s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.958963] env[62569]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 62c1052e-dc0a-4fb6-a399-3e238bf93789 could not be found.", "detail": ""}} {{(pid=62569) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 890.959278] env[62569]: DEBUG nova.network.neutron [-] Unable to show port 62c1052e-dc0a-4fb6-a399-3e238bf93789 as it no longer exists. {{(pid=62569) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 890.991339] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250307, 'name': CreateVM_Task, 'duration_secs': 0.427177} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.991726] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 890.992288] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.992445] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.992785] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 890.993063] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99f6fff7-7426-4cf4-b230-dec70f530018 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.999194] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 890.999194] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52c5c0a9-61d5-8e0e-5656-a29710b2c005" [ 890.999194] env[62569]: _type = "Task" [ 890.999194] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.009495] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52c5c0a9-61d5-8e0e-5656-a29710b2c005, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.027353] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': task-1250308, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.083604] env[62569]: DEBUG nova.scheduler.client.report [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 891.098124] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250294, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.166794] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "refresh_cache-bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.170018] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquired lock "refresh_cache-bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.170018] env[62569]: DEBUG nova.network.neutron [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 891.415749] env[62569]: DEBUG oslo_concurrency.lockutils [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Releasing lock "refresh_cache-31ce29fa-4f60-4404-b830-21ad196f78b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.516025] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52c5c0a9-61d5-8e0e-5656-a29710b2c005, 'name': SearchDatastore_Task, 'duration_secs': 0.02554} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.517756] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.518506] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 891.519354] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.520736] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.520736] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 891.521233] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-95d317be-81e0-4dd3-b3f3-0419abbeaf8c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.532582] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': task-1250308, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.538592] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 891.538592] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 891.538592] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc5cc751-07a0-41b3-b02f-15e0e341eda9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.546966] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 891.546966] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ba91cf-7d65-195d-c54a-4232d13a196e" [ 891.546966] env[62569]: _type = "Task" [ 891.546966] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.560331] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ba91cf-7d65-195d-c54a-4232d13a196e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.563723] env[62569]: DEBUG nova.compute.manager [req-4cf63bc0-7573-4ab6-a0a0-974191d1d4f0 req-02c4ed7f-27a5-40f7-b813-9c45bf655a5a service nova] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Received event network-vif-plugged-d64908ff-b5fa-42fe-8a49-1f39a27ca4b7 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 891.563723] env[62569]: DEBUG oslo_concurrency.lockutils [req-4cf63bc0-7573-4ab6-a0a0-974191d1d4f0 req-02c4ed7f-27a5-40f7-b813-9c45bf655a5a service nova] Acquiring lock "bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.563911] env[62569]: DEBUG oslo_concurrency.lockutils [req-4cf63bc0-7573-4ab6-a0a0-974191d1d4f0 req-02c4ed7f-27a5-40f7-b813-9c45bf655a5a service nova] Lock "bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.565559] env[62569]: DEBUG oslo_concurrency.lockutils [req-4cf63bc0-7573-4ab6-a0a0-974191d1d4f0 req-02c4ed7f-27a5-40f7-b813-9c45bf655a5a service nova] Lock "bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.565851] env[62569]: DEBUG nova.compute.manager [req-4cf63bc0-7573-4ab6-a0a0-974191d1d4f0 req-02c4ed7f-27a5-40f7-b813-9c45bf655a5a service nova] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] No waiting events found dispatching network-vif-plugged-d64908ff-b5fa-42fe-8a49-1f39a27ca4b7 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 891.566163] env[62569]: WARNING nova.compute.manager [req-4cf63bc0-7573-4ab6-a0a0-974191d1d4f0 req-02c4ed7f-27a5-40f7-b813-9c45bf655a5a service nova] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Received unexpected event network-vif-plugged-d64908ff-b5fa-42fe-8a49-1f39a27ca4b7 for instance with vm_state building and task_state spawning. [ 891.566414] env[62569]: DEBUG nova.compute.manager [req-4cf63bc0-7573-4ab6-a0a0-974191d1d4f0 req-02c4ed7f-27a5-40f7-b813-9c45bf655a5a service nova] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Received event network-changed-d64908ff-b5fa-42fe-8a49-1f39a27ca4b7 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 891.566663] env[62569]: DEBUG nova.compute.manager [req-4cf63bc0-7573-4ab6-a0a0-974191d1d4f0 req-02c4ed7f-27a5-40f7-b813-9c45bf655a5a service nova] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Refreshing instance network info cache due to event network-changed-d64908ff-b5fa-42fe-8a49-1f39a27ca4b7. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 891.566956] env[62569]: DEBUG oslo_concurrency.lockutils [req-4cf63bc0-7573-4ab6-a0a0-974191d1d4f0 req-02c4ed7f-27a5-40f7-b813-9c45bf655a5a service nova] Acquiring lock "refresh_cache-bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.593412] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.951s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.603139] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250294, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.621140] env[62569]: INFO nova.scheduler.client.report [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Deleted allocations for instance 298efba9-226d-4105-8b32-76b3dddddb0b [ 891.825064] env[62569]: DEBUG nova.network.neutron [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 891.950480] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4de32a96-e48f-4462-a025-14868cd3684a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.975458] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ca9caf-196a-42a3-8511-583beeb97577 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.985521] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Updating instance '31ce29fa-4f60-4404-b830-21ad196f78b5' progress to 83 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 892.025226] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': task-1250308, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.032740] env[62569]: DEBUG nova.network.neutron [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Updating instance_info_cache with network_info: [{"id": "d64908ff-b5fa-42fe-8a49-1f39a27ca4b7", "address": "fa:16:3e:66:5c:fb", "network": {"id": "334a2e22-0249-48e3-a2f5-ac37df9771d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1048653043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f7fb3e698324f8e851aefbee74ac669", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f880ac2e-d532-4f54-87bb-998a8d1bca78", "external-id": "nsx-vlan-transportzone-491", "segmentation_id": 491, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd64908ff-b5", "ovs_interfaceid": "d64908ff-b5fa-42fe-8a49-1f39a27ca4b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.068457] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ba91cf-7d65-195d-c54a-4232d13a196e, 'name': SearchDatastore_Task, 'duration_secs': 0.013887} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.071039] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8dae5ad-c198-4636-9b6b-2dd993e10978 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.081024] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 892.081024] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52909694-3b13-902d-c857-50588469acb2" [ 892.081024] env[62569]: _type = "Task" [ 892.081024] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.101889] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52909694-3b13-902d-c857-50588469acb2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.119354] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250294, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.769013} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.119973] env[62569]: INFO nova.virt.vmwareapi.ds_util [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_bead562b-e9db-4a1b-b619-67588ba07eae/OSTACK_IMG_bead562b-e9db-4a1b-b619-67588ba07eae.vmdk to [datastore1] devstack-image-cache_base/5909e6a6-6c00-4956-a4d0-c6430d91b165/5909e6a6-6c00-4956-a4d0-c6430d91b165.vmdk. [ 892.119973] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Cleaning up location [datastore1] OSTACK_IMG_bead562b-e9db-4a1b-b619-67588ba07eae {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 892.119973] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_bead562b-e9db-4a1b-b619-67588ba07eae {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 892.120231] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-db08755e-a0b6-44a7-99ec-4d4260a6fe42 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.133995] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d919d637-5d23-44ee-9aa5-3dc53b160550 tempest-InstanceActionsTestJSON-1568033895 tempest-InstanceActionsTestJSON-1568033895-project-member] Lock "298efba9-226d-4105-8b32-76b3dddddb0b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.494s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.136785] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 892.136785] env[62569]: value = "task-1250309" [ 892.136785] env[62569]: _type = "Task" [ 892.136785] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.149538] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250309, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.385915] env[62569]: DEBUG nova.network.neutron [-] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.496622] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 892.496964] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ab1986aa-fd4d-4a41-9348-8f3942b5a63d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.510484] env[62569]: DEBUG oslo_vmware.api [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 892.510484] env[62569]: value = "task-1250310" [ 892.510484] env[62569]: _type = "Task" [ 892.510484] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.523847] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': task-1250308, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.769667} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.527420] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] a6ee3547-3a3d-4480-b451-d1fddd829b39/a6ee3547-3a3d-4480-b451-d1fddd829b39.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 892.527844] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 892.528185] env[62569]: DEBUG oslo_vmware.api [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250310, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.528420] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d3b3107a-3abd-4878-b68b-24628f3d33db {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.535605] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Releasing lock "refresh_cache-bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.535955] env[62569]: DEBUG nova.compute.manager [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Instance network_info: |[{"id": "d64908ff-b5fa-42fe-8a49-1f39a27ca4b7", "address": "fa:16:3e:66:5c:fb", "network": {"id": "334a2e22-0249-48e3-a2f5-ac37df9771d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1048653043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f7fb3e698324f8e851aefbee74ac669", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f880ac2e-d532-4f54-87bb-998a8d1bca78", "external-id": "nsx-vlan-transportzone-491", "segmentation_id": 491, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd64908ff-b5", "ovs_interfaceid": "d64908ff-b5fa-42fe-8a49-1f39a27ca4b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 892.537284] env[62569]: DEBUG oslo_concurrency.lockutils [req-4cf63bc0-7573-4ab6-a0a0-974191d1d4f0 req-02c4ed7f-27a5-40f7-b813-9c45bf655a5a service nova] Acquired lock "refresh_cache-bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.537406] env[62569]: DEBUG nova.network.neutron [req-4cf63bc0-7573-4ab6-a0a0-974191d1d4f0 req-02c4ed7f-27a5-40f7-b813-9c45bf655a5a service nova] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Refreshing network info cache for port d64908ff-b5fa-42fe-8a49-1f39a27ca4b7 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 892.540320] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:5c:fb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f880ac2e-d532-4f54-87bb-998a8d1bca78', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd64908ff-b5fa-42fe-8a49-1f39a27ca4b7', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 892.547250] env[62569]: DEBUG oslo.service.loopingcall [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 892.547423] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Waiting for the task: (returnval){ [ 892.547423] env[62569]: value = "task-1250311" [ 892.547423] env[62569]: _type = "Task" [ 892.547423] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.550813] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 892.551574] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-04139ca2-7da4-4d06-8b16-f603750de873 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.584267] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': task-1250311, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.589936] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 892.589936] env[62569]: value = "task-1250312" [ 892.589936] env[62569]: _type = "Task" [ 892.589936] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.600961] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52909694-3b13-902d-c857-50588469acb2, 'name': SearchDatastore_Task, 'duration_secs': 0.056452} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.600961] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.601143] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 534c9ece-c8df-4528-83d8-c6c0ef8c4793/534c9ece-c8df-4528-83d8-c6c0ef8c4793.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 892.601833] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-33fbe35e-71eb-4f7b-8855-443002a6624c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.607754] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250312, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.613711] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 892.613711] env[62569]: value = "task-1250313" [ 892.613711] env[62569]: _type = "Task" [ 892.613711] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.626787] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250313, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.654578] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250309, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140607} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.654578] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 892.654578] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5909e6a6-6c00-4956-a4d0-c6430d91b165/5909e6a6-6c00-4956-a4d0-c6430d91b165.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.654811] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5909e6a6-6c00-4956-a4d0-c6430d91b165/5909e6a6-6c00-4956-a4d0-c6430d91b165.vmdk to [datastore1] 9c1c0379-3968-464b-b587-8fac704404d5/9c1c0379-3968-464b-b587-8fac704404d5.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 892.654918] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2802fb21-89a6-44a9-b377-4bd732bf5a39 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.665567] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 892.665567] env[62569]: value = "task-1250314" [ 892.665567] env[62569]: _type = "Task" [ 892.665567] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.677297] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250314, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.689483] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b6963b5-bcc3-4f7e-aa82-4e266677d1b7 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Volume attach. Driver type: vmdk {{(pid=62569) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 892.689766] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b6963b5-bcc3-4f7e-aa82-4e266677d1b7 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269461', 'volume_id': 'd20ca906-3332-4f05-bd83-1a0427d2b692', 'name': 'volume-d20ca906-3332-4f05-bd83-1a0427d2b692', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '108d5bde-e463-4a69-816b-bdd3c03e13cd', 'attached_at': '', 'detached_at': '', 'volume_id': 'd20ca906-3332-4f05-bd83-1a0427d2b692', 'serial': 'd20ca906-3332-4f05-bd83-1a0427d2b692'} {{(pid=62569) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 892.690728] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65e4fd19-a9ee-4aa1-8726-ca6448958403 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.714234] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-436268d5-3b61-4a02-8c13-b6d5c3a6a69e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.743506] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b6963b5-bcc3-4f7e-aa82-4e266677d1b7 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] volume-d20ca906-3332-4f05-bd83-1a0427d2b692/volume-d20ca906-3332-4f05-bd83-1a0427d2b692.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.743924] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c32a6cc7-3c63-41a2-bbe4-05101a5bbba5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.765019] env[62569]: DEBUG oslo_vmware.api [None req-5b6963b5-bcc3-4f7e-aa82-4e266677d1b7 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 892.765019] env[62569]: value = "task-1250315" [ 892.765019] env[62569]: _type = "Task" [ 892.765019] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.778148] env[62569]: DEBUG oslo_vmware.api [None req-5b6963b5-bcc3-4f7e-aa82-4e266677d1b7 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250315, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.889173] env[62569]: INFO nova.compute.manager [-] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Took 2.01 seconds to deallocate network for instance. [ 893.028819] env[62569]: DEBUG oslo_vmware.api [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250310, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.068996] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': task-1250311, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076691} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.069296] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 893.074927] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6944e34a-7381-4e7d-960f-3059b14711cc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.110257] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] a6ee3547-3a3d-4480-b451-d1fddd829b39/a6ee3547-3a3d-4480-b451-d1fddd829b39.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.115136] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-834bd63b-5741-4562-9a76-f143f3f0eda3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.133475] env[62569]: DEBUG nova.network.neutron [req-4cf63bc0-7573-4ab6-a0a0-974191d1d4f0 req-02c4ed7f-27a5-40f7-b813-9c45bf655a5a service nova] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Updated VIF entry in instance network info cache for port d64908ff-b5fa-42fe-8a49-1f39a27ca4b7. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 893.134068] env[62569]: DEBUG nova.network.neutron [req-4cf63bc0-7573-4ab6-a0a0-974191d1d4f0 req-02c4ed7f-27a5-40f7-b813-9c45bf655a5a service nova] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Updating instance_info_cache with network_info: [{"id": "d64908ff-b5fa-42fe-8a49-1f39a27ca4b7", "address": "fa:16:3e:66:5c:fb", "network": {"id": "334a2e22-0249-48e3-a2f5-ac37df9771d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1048653043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f7fb3e698324f8e851aefbee74ac669", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f880ac2e-d532-4f54-87bb-998a8d1bca78", "external-id": "nsx-vlan-transportzone-491", "segmentation_id": 491, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd64908ff-b5", "ovs_interfaceid": "d64908ff-b5fa-42fe-8a49-1f39a27ca4b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.155565] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250312, 'name': CreateVM_Task, 'duration_secs': 0.512393} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.162639] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 893.162918] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250313, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.163195] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Waiting for the task: (returnval){ [ 893.163195] env[62569]: value = "task-1250316" [ 893.163195] env[62569]: _type = "Task" [ 893.163195] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.163920] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.164113] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.164495] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 893.165223] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fc90c37-6210-4598-97d6-739840b9cbef {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.182491] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 893.182491] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52c22b1e-22ec-3575-e00b-c76b5a6eb732" [ 893.182491] env[62569]: _type = "Task" [ 893.182491] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.193479] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': task-1250316, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.193829] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250314, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.208199] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52c22b1e-22ec-3575-e00b-c76b5a6eb732, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.278779] env[62569]: DEBUG oslo_vmware.api [None req-5b6963b5-bcc3-4f7e-aa82-4e266677d1b7 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250315, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.401094] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.401639] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.402325] env[62569]: DEBUG nova.objects.instance [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lazy-loading 'resources' on Instance uuid 807d3025-d6a7-4778-a829-a61e2c7495c9 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.458766] env[62569]: DEBUG oslo_concurrency.lockutils [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "a0b406c3-9466-41bd-9de1-e675cab2ceef" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.460403] env[62569]: DEBUG oslo_concurrency.lockutils [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "a0b406c3-9466-41bd-9de1-e675cab2ceef" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.460403] env[62569]: DEBUG oslo_concurrency.lockutils [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "a0b406c3-9466-41bd-9de1-e675cab2ceef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.460403] env[62569]: DEBUG oslo_concurrency.lockutils [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "a0b406c3-9466-41bd-9de1-e675cab2ceef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.460403] env[62569]: DEBUG oslo_concurrency.lockutils [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "a0b406c3-9466-41bd-9de1-e675cab2ceef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.464744] env[62569]: INFO nova.compute.manager [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Terminating instance [ 893.525962] env[62569]: DEBUG oslo_vmware.api [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250310, 'name': PowerOnVM_Task, 'duration_secs': 0.676054} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.526339] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 893.526573] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-609bf980-d12d-4d49-9dfa-a130b682c83d tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Updating instance '31ce29fa-4f60-4404-b830-21ad196f78b5' progress to 100 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 893.616262] env[62569]: DEBUG nova.compute.manager [req-36f0ff22-e20e-46ed-ade0-d1623cf75926 req-b9089656-dd6d-4700-a045-5d6ce6884f3f service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Received event network-vif-deleted-9d855102-33ab-4574-b791-dfa6fb5b0f09 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 893.641430] env[62569]: DEBUG oslo_concurrency.lockutils [req-4cf63bc0-7573-4ab6-a0a0-974191d1d4f0 req-02c4ed7f-27a5-40f7-b813-9c45bf655a5a service nova] Releasing lock "refresh_cache-bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.651296] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250313, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.781439} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.651906] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 534c9ece-c8df-4528-83d8-c6c0ef8c4793/534c9ece-c8df-4528-83d8-c6c0ef8c4793.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 893.652143] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 893.652297] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1693f792-b512-46f8-b59a-793e379e1a6e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.665679] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 893.665679] env[62569]: value = "task-1250317" [ 893.665679] env[62569]: _type = "Task" [ 893.665679] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.687161] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': task-1250316, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.693794] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250314, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.694150] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250317, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.706772] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52c22b1e-22ec-3575-e00b-c76b5a6eb732, 'name': SearchDatastore_Task, 'duration_secs': 0.067897} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.707139] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.707392] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 893.707643] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.708153] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.708153] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 893.708705] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ecac2c3d-4abe-4ccd-8f72-d3e777df2993 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.730400] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 893.730663] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 893.731740] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba7ee87d-2137-4a37-a98b-b93a880bc728 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.740227] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 893.740227] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a56ed7-6ae9-c9bb-4d13-23c9a0d0377d" [ 893.740227] env[62569]: _type = "Task" [ 893.740227] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.752787] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a56ed7-6ae9-c9bb-4d13-23c9a0d0377d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.777684] env[62569]: DEBUG oslo_vmware.api [None req-5b6963b5-bcc3-4f7e-aa82-4e266677d1b7 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250315, 'name': ReconfigVM_Task, 'duration_secs': 0.752282} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.778032] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b6963b5-bcc3-4f7e-aa82-4e266677d1b7 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Reconfigured VM instance instance-00000052 to attach disk [datastore1] volume-d20ca906-3332-4f05-bd83-1a0427d2b692/volume-d20ca906-3332-4f05-bd83-1a0427d2b692.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 893.784620] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8ef781ff-3c31-45c1-ae38-5895067f4bb7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.803501] env[62569]: DEBUG oslo_vmware.api [None req-5b6963b5-bcc3-4f7e-aa82-4e266677d1b7 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 893.803501] env[62569]: value = "task-1250318" [ 893.803501] env[62569]: _type = "Task" [ 893.803501] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.814304] env[62569]: DEBUG oslo_vmware.api [None req-5b6963b5-bcc3-4f7e-aa82-4e266677d1b7 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250318, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.970113] env[62569]: DEBUG nova.compute.manager [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 893.970615] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 893.971842] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13edccd5-0224-44fe-ac74-ab982cef188f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.989374] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 893.992870] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4b10eb0d-5958-4036-b61b-d77ab76fe028 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.006640] env[62569]: DEBUG oslo_vmware.api [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 894.006640] env[62569]: value = "task-1250319" [ 894.006640] env[62569]: _type = "Task" [ 894.006640] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.022205] env[62569]: DEBUG oslo_vmware.api [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250319, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.188669] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': task-1250316, 'name': ReconfigVM_Task, 'duration_secs': 0.658795} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.193240] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Reconfigured VM instance instance-00000054 to attach disk [datastore1] a6ee3547-3a3d-4480-b451-d1fddd829b39/a6ee3547-3a3d-4480-b451-d1fddd829b39.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.194584] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250317, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.194949] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8f467f66-beea-4cef-a605-a2607788ff45 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.202784] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250314, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.208573] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Waiting for the task: (returnval){ [ 894.208573] env[62569]: value = "task-1250320" [ 894.208573] env[62569]: _type = "Task" [ 894.208573] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.223973] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': task-1250320, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.256563] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a56ed7-6ae9-c9bb-4d13-23c9a0d0377d, 'name': SearchDatastore_Task, 'duration_secs': 0.103791} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.257596] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21c15708-61ed-4d77-b38d-54ee7aeb9db4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.267625] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 894.267625] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5295481b-c7d6-d7ca-4e9b-8315ec787146" [ 894.267625] env[62569]: _type = "Task" [ 894.267625] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.289789] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5295481b-c7d6-d7ca-4e9b-8315ec787146, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.304408] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aba2128b-4bf5-4190-a3cf-1e9153aa93bf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.325044] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbbaa25b-3837-48e9-9421-b76754bd7a47 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.331047] env[62569]: DEBUG oslo_vmware.api [None req-5b6963b5-bcc3-4f7e-aa82-4e266677d1b7 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250318, 'name': ReconfigVM_Task, 'duration_secs': 0.29211} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.331424] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b6963b5-bcc3-4f7e-aa82-4e266677d1b7 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269461', 'volume_id': 'd20ca906-3332-4f05-bd83-1a0427d2b692', 'name': 'volume-d20ca906-3332-4f05-bd83-1a0427d2b692', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '108d5bde-e463-4a69-816b-bdd3c03e13cd', 'attached_at': '', 'detached_at': '', 'volume_id': 'd20ca906-3332-4f05-bd83-1a0427d2b692', 'serial': 'd20ca906-3332-4f05-bd83-1a0427d2b692'} {{(pid=62569) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 894.379155] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7deb0a3b-fd30-4d35-bcac-89c964f5c3d3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.392575] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31b7e0d7-ea61-4ce6-88a3-7b407b7e1538 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.414678] env[62569]: DEBUG nova.compute.provider_tree [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.523689] env[62569]: DEBUG oslo_vmware.api [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250319, 'name': PowerOffVM_Task, 'duration_secs': 0.414056} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.524268] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 894.524531] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 894.524833] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-02b65192-c47d-4e9c-864d-261e2dab5fd0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.628672] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 894.629278] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 894.629636] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Deleting the datastore file [datastore2] a0b406c3-9466-41bd-9de1-e675cab2ceef {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 894.629963] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c700808d-7baf-45db-90f5-66194bdf23c2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.649614] env[62569]: DEBUG oslo_vmware.api [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 894.649614] env[62569]: value = "task-1250322" [ 894.649614] env[62569]: _type = "Task" [ 894.649614] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.667496] env[62569]: DEBUG oslo_vmware.api [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250322, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.688798] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250317, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.826691} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.692826] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 894.692826] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250314, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.693427] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff99a693-ee9a-40e5-8886-9e0341961d7a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.718869] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] 534c9ece-c8df-4528-83d8-c6c0ef8c4793/534c9ece-c8df-4528-83d8-c6c0ef8c4793.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 894.723226] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1f724b7-3d49-4350-8d3f-7069615ce0a5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.746763] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': task-1250320, 'name': Rename_Task, 'duration_secs': 0.281508} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.748530] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 894.748898] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 894.748898] env[62569]: value = "task-1250323" [ 894.748898] env[62569]: _type = "Task" [ 894.748898] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.749269] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2a4aaf4a-35dd-464b-a129-82a107316a22 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.762821] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250323, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.764551] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Waiting for the task: (returnval){ [ 894.764551] env[62569]: value = "task-1250324" [ 894.764551] env[62569]: _type = "Task" [ 894.764551] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.787263] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5295481b-c7d6-d7ca-4e9b-8315ec787146, 'name': SearchDatastore_Task, 'duration_secs': 0.111198} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.788397] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.790862] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a/bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 894.790862] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': task-1250324, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.790862] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a94542fb-6655-4413-893c-8d5d4c376584 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.801703] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 894.801703] env[62569]: value = "task-1250325" [ 894.801703] env[62569]: _type = "Task" [ 894.801703] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.818153] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250325, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.919065] env[62569]: DEBUG nova.scheduler.client.report [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 895.164548] env[62569]: DEBUG oslo_vmware.api [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250322, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.194224] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250314, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.264256] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250323, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.278915] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': task-1250324, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.315324] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250325, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.419641] env[62569]: DEBUG nova.objects.instance [None req-5b6963b5-bcc3-4f7e-aa82-4e266677d1b7 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lazy-loading 'flavor' on Instance uuid 108d5bde-e463-4a69-816b-bdd3c03e13cd {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 895.426939] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.025s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.455143] env[62569]: INFO nova.scheduler.client.report [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Deleted allocations for instance 807d3025-d6a7-4778-a829-a61e2c7495c9 [ 895.664744] env[62569]: DEBUG oslo_vmware.api [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250322, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.524971} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.665060] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 895.665284] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 895.665504] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 895.665696] env[62569]: INFO nova.compute.manager [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Took 1.70 seconds to destroy the instance on the hypervisor. [ 895.665953] env[62569]: DEBUG oslo.service.loopingcall [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.666185] env[62569]: DEBUG nova.compute.manager [-] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 895.666339] env[62569]: DEBUG nova.network.neutron [-] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 895.695336] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250314, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.770946] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250323, 'name': ReconfigVM_Task, 'duration_secs': 0.519372} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.774628] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Reconfigured VM instance instance-00000055 to attach disk [datastore1] 534c9ece-c8df-4528-83d8-c6c0ef8c4793/534c9ece-c8df-4528-83d8-c6c0ef8c4793.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 895.775360] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-403716e3-fcf4-4b82-b0d1-cb2dbea17d30 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.784255] env[62569]: DEBUG oslo_vmware.api [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': task-1250324, 'name': PowerOnVM_Task, 'duration_secs': 0.806126} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.785595] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 895.785819] env[62569]: INFO nova.compute.manager [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Took 11.17 seconds to spawn the instance on the hypervisor. [ 895.786058] env[62569]: DEBUG nova.compute.manager [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 895.786775] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 895.786775] env[62569]: value = "task-1250326" [ 895.786775] env[62569]: _type = "Task" [ 895.786775] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.787639] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f628a8ef-dbd4-46d8-9b23-d6ca8b64e0da {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.798645] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250326, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.812486] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250325, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.864394] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "108d5bde-e463-4a69-816b-bdd3c03e13cd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.932749] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5b6963b5-bcc3-4f7e-aa82-4e266677d1b7 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "108d5bde-e463-4a69-816b-bdd3c03e13cd" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.864s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.934115] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "108d5bde-e463-4a69-816b-bdd3c03e13cd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.070s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.935443] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "108d5bde-e463-4a69-816b-bdd3c03e13cd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.935443] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "108d5bde-e463-4a69-816b-bdd3c03e13cd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.935443] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "108d5bde-e463-4a69-816b-bdd3c03e13cd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.937694] env[62569]: INFO nova.compute.manager [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Terminating instance [ 895.965895] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e1b3dc42-4453-4edc-9147-eb1f82b8f3b1 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "807d3025-d6a7-4778-a829-a61e2c7495c9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.830s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.969242] env[62569]: DEBUG oslo_concurrency.lockutils [req-7affada8-b78d-45ff-be6a-e5d7fca572e2 req-27a73008-78ff-48fa-b162-fe3fca2d1efc service nova] Acquired lock "807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.970481] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93948518-2e35-4343-87f0-db5842a49737 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.980204] env[62569]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 895.980336] env[62569]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=62569) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 895.980846] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e65bddf0-b50f-49ff-8c2f-1e1f05f0ab06 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.992090] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83d441ca-08d0-4ecc-8ac6-4e79a9c08933 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.032482] env[62569]: ERROR root [req-7affada8-b78d-45ff-be6a-e5d7fca572e2 req-27a73008-78ff-48fa-b162-fe3fca2d1efc service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-269408' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-269408' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-269408' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-269408'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-269408' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-269408' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-269408'}\n"]: nova.exception.InstanceNotFound: Instance 807d3025-d6a7-4778-a829-a61e2c7495c9 could not be found. [ 896.032996] env[62569]: DEBUG oslo_concurrency.lockutils [req-7affada8-b78d-45ff-be6a-e5d7fca572e2 req-27a73008-78ff-48fa-b162-fe3fca2d1efc service nova] Releasing lock "807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.033061] env[62569]: DEBUG nova.compute.manager [req-7affada8-b78d-45ff-be6a-e5d7fca572e2 req-27a73008-78ff-48fa-b162-fe3fca2d1efc service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Detach interface failed, port_id=5f2ad2ec-3e95-411c-ac38-94f6d0e8aa18, reason: Instance 807d3025-d6a7-4778-a829-a61e2c7495c9 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 896.034025] env[62569]: DEBUG nova.compute.manager [req-7affada8-b78d-45ff-be6a-e5d7fca572e2 req-27a73008-78ff-48fa-b162-fe3fca2d1efc service nova] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Received event network-vif-plugged-37c7a902-07de-45dc-b683-311b84a5edc5 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 896.034025] env[62569]: DEBUG oslo_concurrency.lockutils [req-7affada8-b78d-45ff-be6a-e5d7fca572e2 req-27a73008-78ff-48fa-b162-fe3fca2d1efc service nova] Acquiring lock "534c9ece-c8df-4528-83d8-c6c0ef8c4793-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.034025] env[62569]: DEBUG oslo_concurrency.lockutils [req-7affada8-b78d-45ff-be6a-e5d7fca572e2 req-27a73008-78ff-48fa-b162-fe3fca2d1efc service nova] Lock "534c9ece-c8df-4528-83d8-c6c0ef8c4793-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.034025] env[62569]: DEBUG oslo_concurrency.lockutils [req-7affada8-b78d-45ff-be6a-e5d7fca572e2 req-27a73008-78ff-48fa-b162-fe3fca2d1efc service nova] Lock "534c9ece-c8df-4528-83d8-c6c0ef8c4793-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.034363] env[62569]: DEBUG nova.compute.manager [req-7affada8-b78d-45ff-be6a-e5d7fca572e2 req-27a73008-78ff-48fa-b162-fe3fca2d1efc service nova] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] No waiting events found dispatching network-vif-plugged-37c7a902-07de-45dc-b683-311b84a5edc5 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 896.034601] env[62569]: WARNING nova.compute.manager [req-7affada8-b78d-45ff-be6a-e5d7fca572e2 req-27a73008-78ff-48fa-b162-fe3fca2d1efc service nova] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Received unexpected event network-vif-plugged-37c7a902-07de-45dc-b683-311b84a5edc5 for instance with vm_state building and task_state spawning. [ 896.034818] env[62569]: DEBUG nova.compute.manager [req-7affada8-b78d-45ff-be6a-e5d7fca572e2 req-27a73008-78ff-48fa-b162-fe3fca2d1efc service nova] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Received event network-changed-37c7a902-07de-45dc-b683-311b84a5edc5 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 896.035146] env[62569]: DEBUG nova.compute.manager [req-7affada8-b78d-45ff-be6a-e5d7fca572e2 req-27a73008-78ff-48fa-b162-fe3fca2d1efc service nova] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Refreshing instance network info cache due to event network-changed-37c7a902-07de-45dc-b683-311b84a5edc5. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 896.035319] env[62569]: DEBUG oslo_concurrency.lockutils [req-7affada8-b78d-45ff-be6a-e5d7fca572e2 req-27a73008-78ff-48fa-b162-fe3fca2d1efc service nova] Acquiring lock "refresh_cache-534c9ece-c8df-4528-83d8-c6c0ef8c4793" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.035519] env[62569]: DEBUG oslo_concurrency.lockutils [req-7affada8-b78d-45ff-be6a-e5d7fca572e2 req-27a73008-78ff-48fa-b162-fe3fca2d1efc service nova] Acquired lock "refresh_cache-534c9ece-c8df-4528-83d8-c6c0ef8c4793" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.035749] env[62569]: DEBUG nova.network.neutron [req-7affada8-b78d-45ff-be6a-e5d7fca572e2 req-27a73008-78ff-48fa-b162-fe3fca2d1efc service nova] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Refreshing network info cache for port 37c7a902-07de-45dc-b683-311b84a5edc5 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 896.037694] env[62569]: DEBUG oslo_concurrency.lockutils [req-2ba108f8-5e2b-4a70-8613-12bbca19a584 req-0973b397-db6e-4d35-87d6-7273a151ce44 service nova] Acquired lock "807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.038618] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deebd91f-1d94-4052-9ec6-4e719816179c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.049118] env[62569]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 896.049292] env[62569]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=62569) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 896.049762] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e4bba0c6-daa2-4f36-9691-54d2007bba8a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.063153] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e310ee6c-cc1a-4ab4-a1aa-5c94a5e25495 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.105676] env[62569]: ERROR root [req-2ba108f8-5e2b-4a70-8613-12bbca19a584 req-0973b397-db6e-4d35-87d6-7273a151ce44 service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-269408' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-269408' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-269408' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-269408'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-269408' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-269408' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-269408'}\n"]: nova.exception.InstanceNotFound: Instance 807d3025-d6a7-4778-a829-a61e2c7495c9 could not be found. [ 896.105928] env[62569]: DEBUG oslo_concurrency.lockutils [req-2ba108f8-5e2b-4a70-8613-12bbca19a584 req-0973b397-db6e-4d35-87d6-7273a151ce44 service nova] Releasing lock "807d3025-d6a7-4778-a829-a61e2c7495c9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.107346] env[62569]: DEBUG nova.compute.manager [req-2ba108f8-5e2b-4a70-8613-12bbca19a584 req-0973b397-db6e-4d35-87d6-7273a151ce44 service nova] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Detach interface failed, port_id=62c1052e-dc0a-4fb6-a399-3e238bf93789, reason: Instance 807d3025-d6a7-4778-a829-a61e2c7495c9 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 896.144939] env[62569]: DEBUG oslo_concurrency.lockutils [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "b6e5eefc-8c06-445b-a3af-9404578b6179" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.145250] env[62569]: DEBUG oslo_concurrency.lockutils [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "b6e5eefc-8c06-445b-a3af-9404578b6179" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.145441] env[62569]: INFO nova.compute.manager [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Shelving [ 896.194746] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250314, 'name': CopyVirtualDisk_Task, 'duration_secs': 3.158189} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.195115] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5909e6a6-6c00-4956-a4d0-c6430d91b165/5909e6a6-6c00-4956-a4d0-c6430d91b165.vmdk to [datastore1] 9c1c0379-3968-464b-b587-8fac704404d5/9c1c0379-3968-464b-b587-8fac704404d5.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 896.195937] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ddf0837-4a24-478a-a034-06f4151ef366 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.225550] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 9c1c0379-3968-464b-b587-8fac704404d5/9c1c0379-3968-464b-b587-8fac704404d5.vmdk or device None with type streamOptimized {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 896.226897] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4abb0f23-a29e-495e-af7b-4d41bb10705f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.259052] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 896.259052] env[62569]: value = "task-1250327" [ 896.259052] env[62569]: _type = "Task" [ 896.259052] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.272022] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250327, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.301400] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250326, 'name': Rename_Task, 'duration_secs': 0.1849} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.301744] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 896.302112] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bac48671-4a2c-45f8-98a4-1cff12c5e0cc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.314904] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 896.314904] env[62569]: value = "task-1250328" [ 896.314904] env[62569]: _type = "Task" [ 896.314904] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.319108] env[62569]: INFO nova.compute.manager [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Took 23.65 seconds to build instance. [ 896.323932] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250325, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.499992} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.329444] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a/bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 896.329684] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 896.329970] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0249d646-a0b9-49a0-84e9-90d783b41fd3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.338747] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250328, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.340789] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 896.340789] env[62569]: value = "task-1250329" [ 896.340789] env[62569]: _type = "Task" [ 896.340789] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.350251] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250329, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.442801] env[62569]: DEBUG nova.compute.manager [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 896.443169] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 896.443488] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-81f3361f-e09c-4a11-b248-c3d58d554a21 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.453789] env[62569]: DEBUG oslo_vmware.api [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 896.453789] env[62569]: value = "task-1250330" [ 896.453789] env[62569]: _type = "Task" [ 896.453789] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.468539] env[62569]: DEBUG oslo_vmware.api [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250330, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.503520] env[62569]: DEBUG nova.compute.manager [req-31c02be2-327f-41ac-a7fe-4cf285e3d643 req-3c921bf9-6e0d-448c-863f-945b1c0ad09e service nova] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Received event network-vif-deleted-3da67b46-e89d-4a96-8dd9-752956b9b270 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 896.503767] env[62569]: INFO nova.compute.manager [req-31c02be2-327f-41ac-a7fe-4cf285e3d643 req-3c921bf9-6e0d-448c-863f-945b1c0ad09e service nova] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Neutron deleted interface 3da67b46-e89d-4a96-8dd9-752956b9b270; detaching it from the instance and deleting it from the info cache [ 896.503983] env[62569]: DEBUG nova.network.neutron [req-31c02be2-327f-41ac-a7fe-4cf285e3d643 req-3c921bf9-6e0d-448c-863f-945b1c0ad09e service nova] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.706381] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f6f79371-bbdb-4d19-bd3b-66c4ed4687fe tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "31ce29fa-4f60-4404-b830-21ad196f78b5" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.706671] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f6f79371-bbdb-4d19-bd3b-66c4ed4687fe tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "31ce29fa-4f60-4404-b830-21ad196f78b5" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.706889] env[62569]: DEBUG nova.compute.manager [None req-f6f79371-bbdb-4d19-bd3b-66c4ed4687fe tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Going to confirm migration 1 {{(pid=62569) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5156}} [ 896.771592] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250327, 'name': ReconfigVM_Task, 'duration_secs': 0.43678} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.771592] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 9c1c0379-3968-464b-b587-8fac704404d5/9c1c0379-3968-464b-b587-8fac704404d5.vmdk or device None with type streamOptimized {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 896.772435] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-77287b61-ae8f-4bbe-87f9-d53043cff478 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.774418] env[62569]: DEBUG nova.network.neutron [-] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.781957] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 896.781957] env[62569]: value = "task-1250331" [ 896.781957] env[62569]: _type = "Task" [ 896.781957] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.803022] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250331, 'name': Rename_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.825058] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02884a85-a740-42bc-b8f2-72a2d643b5cf tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Lock "a6ee3547-3a3d-4480-b451-d1fddd829b39" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.182s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.837969] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250328, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.854636] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250329, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078691} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.854636] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 896.854636] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b7d5a0e-7c75-42ea-91f5-6acd1a09b905 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.883956] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a/bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 896.884267] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-806d1a88-d8ad-4d42-9947-05bff89f0f88 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.912692] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 896.912692] env[62569]: value = "task-1250332" [ 896.912692] env[62569]: _type = "Task" [ 896.912692] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.925743] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250332, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.966258] env[62569]: DEBUG oslo_vmware.api [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250330, 'name': PowerOffVM_Task, 'duration_secs': 0.405944} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.966593] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 896.966809] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Volume detach. Driver type: vmdk {{(pid=62569) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 896.967733] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269461', 'volume_id': 'd20ca906-3332-4f05-bd83-1a0427d2b692', 'name': 'volume-d20ca906-3332-4f05-bd83-1a0427d2b692', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '108d5bde-e463-4a69-816b-bdd3c03e13cd', 'attached_at': '', 'detached_at': '', 'volume_id': 'd20ca906-3332-4f05-bd83-1a0427d2b692', 'serial': 'd20ca906-3332-4f05-bd83-1a0427d2b692'} {{(pid=62569) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 896.967895] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c60fc60e-2c72-4042-b695-cda054be485d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.995070] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-428ac971-b0aa-4e60-9beb-948a8a8d1dc6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.006831] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3c0047fa-dcb1-4d9f-aac0-1d6e268a2f39 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.010387] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56eed43c-d01c-4311-8d0b-9f706064048e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.015044] env[62569]: DEBUG nova.network.neutron [req-7affada8-b78d-45ff-be6a-e5d7fca572e2 req-27a73008-78ff-48fa-b162-fe3fca2d1efc service nova] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Updated VIF entry in instance network info cache for port 37c7a902-07de-45dc-b683-311b84a5edc5. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 897.015533] env[62569]: DEBUG nova.network.neutron [req-7affada8-b78d-45ff-be6a-e5d7fca572e2 req-27a73008-78ff-48fa-b162-fe3fca2d1efc service nova] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Updating instance_info_cache with network_info: [{"id": "37c7a902-07de-45dc-b683-311b84a5edc5", "address": "fa:16:3e:e3:66:93", "network": {"id": "884d3248-07f1-4cf7-8ceb-b94b34622384", "bridge": "br-int", "label": "tempest-ServersTestJSON-1882023778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2c6a4814c234d53bbd4844cd482fbae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37c7a902-07", "ovs_interfaceid": "37c7a902-07de-45dc-b683-311b84a5edc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.042904] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1290d8ec-6226-492b-8fab-65429c736732 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.050461] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94f9e233-a35d-406f-93e9-db78cb684672 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.079777] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] The volume has not been displaced from its original location: [datastore1] volume-d20ca906-3332-4f05-bd83-1a0427d2b692/volume-d20ca906-3332-4f05-bd83-1a0427d2b692.vmdk. No consolidation needed. {{(pid=62569) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 897.085793] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Reconfiguring VM instance instance-00000052 to detach disk 2001 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 897.096823] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7b63dd03-a29f-4374-839c-d86e6ec47b4c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.111796] env[62569]: DEBUG nova.compute.manager [req-31c02be2-327f-41ac-a7fe-4cf285e3d643 req-3c921bf9-6e0d-448c-863f-945b1c0ad09e service nova] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Detach interface failed, port_id=3da67b46-e89d-4a96-8dd9-752956b9b270, reason: Instance a0b406c3-9466-41bd-9de1-e675cab2ceef could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 897.120837] env[62569]: DEBUG oslo_vmware.api [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 897.120837] env[62569]: value = "task-1250333" [ 897.120837] env[62569]: _type = "Task" [ 897.120837] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.131122] env[62569]: DEBUG oslo_vmware.api [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250333, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.160131] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 897.160675] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2c01c3ac-8ced-4c6e-9640-cc78e25a5b90 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.169360] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 897.169360] env[62569]: value = "task-1250334" [ 897.169360] env[62569]: _type = "Task" [ 897.169360] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.179183] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250334, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.277873] env[62569]: INFO nova.compute.manager [-] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Took 1.61 seconds to deallocate network for instance. [ 897.296996] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250331, 'name': Rename_Task, 'duration_secs': 0.25755} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.297334] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 897.297600] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3ec8415e-5c50-4967-85f7-7ecd779b9794 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.302476] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f6f79371-bbdb-4d19-bd3b-66c4ed4687fe tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "refresh_cache-31ce29fa-4f60-4404-b830-21ad196f78b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.302476] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f6f79371-bbdb-4d19-bd3b-66c4ed4687fe tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired lock "refresh_cache-31ce29fa-4f60-4404-b830-21ad196f78b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.302476] env[62569]: DEBUG nova.network.neutron [None req-f6f79371-bbdb-4d19-bd3b-66c4ed4687fe tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 897.302476] env[62569]: DEBUG nova.objects.instance [None req-f6f79371-bbdb-4d19-bd3b-66c4ed4687fe tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lazy-loading 'info_cache' on Instance uuid 31ce29fa-4f60-4404-b830-21ad196f78b5 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.310050] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 897.310050] env[62569]: value = "task-1250335" [ 897.310050] env[62569]: _type = "Task" [ 897.310050] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.320654] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250335, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.337200] env[62569]: DEBUG oslo_vmware.api [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250328, 'name': PowerOnVM_Task, 'duration_secs': 0.658708} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.337541] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 897.337722] env[62569]: INFO nova.compute.manager [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Took 10.22 seconds to spawn the instance on the hypervisor. [ 897.337920] env[62569]: DEBUG nova.compute.manager [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 897.339049] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d75d312-ec2f-4cd2-9d56-13a634949aa2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.425822] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250332, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.520331] env[62569]: DEBUG oslo_concurrency.lockutils [req-7affada8-b78d-45ff-be6a-e5d7fca572e2 req-27a73008-78ff-48fa-b162-fe3fca2d1efc service nova] Releasing lock "refresh_cache-534c9ece-c8df-4528-83d8-c6c0ef8c4793" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.632804] env[62569]: DEBUG oslo_vmware.api [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250333, 'name': ReconfigVM_Task, 'duration_secs': 0.38989} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.633277] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Reconfigured VM instance instance-00000052 to detach disk 2001 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 897.640622] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e2e0750-e8b0-41e9-93ed-09fcec4f2935 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.661958] env[62569]: DEBUG oslo_vmware.api [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 897.661958] env[62569]: value = "task-1250336" [ 897.661958] env[62569]: _type = "Task" [ 897.661958] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.676486] env[62569]: DEBUG oslo_vmware.api [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250336, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.688478] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250334, 'name': PowerOffVM_Task, 'duration_secs': 0.278409} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.688939] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 897.690523] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eeb9fe4-2788-4eb3-9a30-c20c7d9b7bef {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.712169] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9c8a5e2-7d54-4fd6-a549-de47ae024734 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.791767] env[62569]: DEBUG oslo_concurrency.lockutils [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.792034] env[62569]: DEBUG oslo_concurrency.lockutils [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.792290] env[62569]: DEBUG nova.objects.instance [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lazy-loading 'resources' on Instance uuid a0b406c3-9466-41bd-9de1-e675cab2ceef {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.822591] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250335, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.859799] env[62569]: INFO nova.compute.manager [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Took 20.01 seconds to build instance. [ 897.927527] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250332, 'name': ReconfigVM_Task, 'duration_secs': 0.633091} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.927804] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Reconfigured VM instance instance-00000056 to attach disk [datastore1] bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a/bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 897.928557] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9d6491fc-f865-4b65-bc91-6414f24589c0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.937803] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 897.937803] env[62569]: value = "task-1250337" [ 897.937803] env[62569]: _type = "Task" [ 897.937803] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.950903] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250337, 'name': Rename_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.178734] env[62569]: DEBUG oslo_vmware.api [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250336, 'name': ReconfigVM_Task, 'duration_secs': 0.299113} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.179389] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269461', 'volume_id': 'd20ca906-3332-4f05-bd83-1a0427d2b692', 'name': 'volume-d20ca906-3332-4f05-bd83-1a0427d2b692', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '108d5bde-e463-4a69-816b-bdd3c03e13cd', 'attached_at': '', 'detached_at': '', 'volume_id': 'd20ca906-3332-4f05-bd83-1a0427d2b692', 'serial': 'd20ca906-3332-4f05-bd83-1a0427d2b692'} {{(pid=62569) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 898.180895] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 898.182629] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d6d4d89-cc10-4213-bcfc-5463a4583d2a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.193895] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 898.194535] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-09318782-e42a-463b-8f92-2f9afea50f31 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.225545] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Creating Snapshot of the VM instance {{(pid=62569) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 898.226083] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e9d96286-b37a-48b7-b814-3352a00646af {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.239223] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 898.239223] env[62569]: value = "task-1250339" [ 898.239223] env[62569]: _type = "Task" [ 898.239223] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.255311] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250339, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.281358] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 898.281358] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 898.281358] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Deleting the datastore file [datastore2] 108d5bde-e463-4a69-816b-bdd3c03e13cd {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 898.281358] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7c692e49-5961-4f78-ab10-ce7a44642741 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.292189] env[62569]: DEBUG oslo_vmware.api [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 898.292189] env[62569]: value = "task-1250340" [ 898.292189] env[62569]: _type = "Task" [ 898.292189] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.310666] env[62569]: DEBUG oslo_vmware.api [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250340, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.321358] env[62569]: DEBUG oslo_vmware.api [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250335, 'name': PowerOnVM_Task, 'duration_secs': 0.874507} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.322170] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 898.322170] env[62569]: INFO nova.compute.manager [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Took 16.25 seconds to spawn the instance on the hypervisor. [ 898.322170] env[62569]: DEBUG nova.compute.manager [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 898.322982] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a162dfb-28a0-4766-a2d4-8e3201ce5d36 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.362032] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bcd3dee3-6fca-4d64-85b6-5f33ed9e0bea tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "534c9ece-c8df-4528-83d8-c6c0ef8c4793" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.523s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.448201] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250337, 'name': Rename_Task, 'duration_secs': 0.232318} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.448539] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 898.448821] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d7c5339b-d282-4572-88ef-73c35a93f339 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.459665] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 898.459665] env[62569]: value = "task-1250341" [ 898.459665] env[62569]: _type = "Task" [ 898.459665] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.473274] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250341, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.540782] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c84f6797-1f78-44f7-8ee6-3030a114dfed {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.549932] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e46cbf1-28f2-47ec-8150-fad752dac77a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.591377] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bcfbdf7-4282-4e82-894e-5e8fd0d112ca {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.602612] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0cb51da-9c73-4a96-92ed-bff96653fd1a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.617076] env[62569]: DEBUG nova.compute.provider_tree [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 898.725628] env[62569]: DEBUG nova.network.neutron [None req-f6f79371-bbdb-4d19-bd3b-66c4ed4687fe tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Updating instance_info_cache with network_info: [{"id": "8e55dca6-f777-4333-9b49-feb5527d9a5e", "address": "fa:16:3e:97:39:72", "network": {"id": "c61b7867-7897-43ae-8f5b-19834c4390ea", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-512066189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ba7e96d226942bd99ab29fe703c6fea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e55dca6-f7", "ovs_interfaceid": "8e55dca6-f777-4333-9b49-feb5527d9a5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.752696] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250339, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.804071] env[62569]: DEBUG oslo_vmware.api [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250340, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.304712} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.804071] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 898.804071] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 898.804229] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 898.804265] env[62569]: INFO nova.compute.manager [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Took 2.36 seconds to destroy the instance on the hypervisor. [ 898.804512] env[62569]: DEBUG oslo.service.loopingcall [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 898.804922] env[62569]: DEBUG nova.compute.manager [-] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 898.804922] env[62569]: DEBUG nova.network.neutron [-] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 898.847660] env[62569]: INFO nova.compute.manager [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Took 28.67 seconds to build instance. [ 898.974463] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250341, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.152537] env[62569]: ERROR nova.scheduler.client.report [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [req-a52fa83c-f645-44f2-9768-274019dc67e8] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID fa06556a-5785-4014-b8bd-bc240a0cf716. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a52fa83c-f645-44f2-9768-274019dc67e8"}]} [ 899.173171] env[62569]: DEBUG nova.scheduler.client.report [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Refreshing inventories for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 899.197266] env[62569]: DEBUG nova.scheduler.client.report [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Updating ProviderTree inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 899.197526] env[62569]: DEBUG nova.compute.provider_tree [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 899.204117] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Acquiring lock "a6ee3547-3a3d-4480-b451-d1fddd829b39" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.204503] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Lock "a6ee3547-3a3d-4480-b451-d1fddd829b39" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.205100] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Acquiring lock "a6ee3547-3a3d-4480-b451-d1fddd829b39-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.205100] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Lock "a6ee3547-3a3d-4480-b451-d1fddd829b39-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.205286] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Lock "a6ee3547-3a3d-4480-b451-d1fddd829b39-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.211261] env[62569]: INFO nova.compute.manager [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Terminating instance [ 899.219238] env[62569]: DEBUG nova.scheduler.client.report [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Refreshing aggregate associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, aggregates: None {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 899.232328] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f6f79371-bbdb-4d19-bd3b-66c4ed4687fe tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Releasing lock "refresh_cache-31ce29fa-4f60-4404-b830-21ad196f78b5" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.232645] env[62569]: DEBUG nova.objects.instance [None req-f6f79371-bbdb-4d19-bd3b-66c4ed4687fe tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lazy-loading 'migration_context' on Instance uuid 31ce29fa-4f60-4404-b830-21ad196f78b5 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 899.245757] env[62569]: DEBUG nova.scheduler.client.report [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Refreshing trait associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 899.254944] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250339, 'name': CreateSnapshot_Task, 'duration_secs': 0.954728} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.255347] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Created Snapshot of the VM instance {{(pid=62569) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 899.257481] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0958b6cf-18aa-43bf-9cc8-c5cd729df33e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.300196] env[62569]: DEBUG oslo_concurrency.lockutils [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "188cac03-4034-4a02-973a-fb1906399fd1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.300472] env[62569]: DEBUG oslo_concurrency.lockutils [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "188cac03-4034-4a02-973a-fb1906399fd1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.349376] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f8c613b4-555e-4d12-b877-c950631fdad7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "9c1c0379-3968-464b-b587-8fac704404d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.178s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.478515] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250341, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.535121] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-350dbc6c-d731-4117-816c-6bff30a2a60f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.545461] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd010f2a-2915-46cd-8942-32f10788cdb1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.584292] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c52d596c-9ff1-46dc-b76c-9a0eccd2f144 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.597425] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-380e1ae1-a530-4a30-8b1a-875cc56628d6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.618244] env[62569]: DEBUG nova.compute.provider_tree [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 899.623859] env[62569]: DEBUG nova.compute.manager [req-46ede12a-4307-43cc-8b50-bddd84edb065 req-c91ae0d2-69dc-4786-a6d4-85c428eb3df0 service nova] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Received event network-vif-deleted-19f8c74e-ebd3-4b14-9e7b-1539ce706834 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 899.624105] env[62569]: INFO nova.compute.manager [req-46ede12a-4307-43cc-8b50-bddd84edb065 req-c91ae0d2-69dc-4786-a6d4-85c428eb3df0 service nova] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Neutron deleted interface 19f8c74e-ebd3-4b14-9e7b-1539ce706834; detaching it from the instance and deleting it from the info cache [ 899.624541] env[62569]: DEBUG nova.network.neutron [req-46ede12a-4307-43cc-8b50-bddd84edb065 req-c91ae0d2-69dc-4786-a6d4-85c428eb3df0 service nova] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.725045] env[62569]: DEBUG nova.compute.manager [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 899.725045] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 899.725402] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b57e94f-ce70-4a25-add1-f0b7292cc5bf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.736789] env[62569]: DEBUG nova.objects.base [None req-f6f79371-bbdb-4d19-bd3b-66c4ed4687fe tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Object Instance<31ce29fa-4f60-4404-b830-21ad196f78b5> lazy-loaded attributes: info_cache,migration_context {{(pid=62569) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 899.736862] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 899.737764] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e07786-bb40-4656-88e5-c2144cba62bb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.743038] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7aa7abb5-7c4c-4d0f-8ce4-04481007810e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.767640] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce4ef7f5-e4bd-41b4-9dac-35cb1feab90d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.770570] env[62569]: DEBUG oslo_vmware.api [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Waiting for the task: (returnval){ [ 899.770570] env[62569]: value = "task-1250342" [ 899.770570] env[62569]: _type = "Task" [ 899.770570] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.782767] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Creating linked-clone VM from snapshot {{(pid=62569) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 899.783753] env[62569]: DEBUG oslo_vmware.api [None req-f6f79371-bbdb-4d19-bd3b-66c4ed4687fe tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 899.783753] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a18a03-791b-dbb5-0e5d-90ebe372436e" [ 899.783753] env[62569]: _type = "Task" [ 899.783753] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.784009] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ffed7ec4-77c3-417b-b5f9-17f1a3cee773 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.795097] env[62569]: DEBUG oslo_vmware.api [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': task-1250342, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.804170] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 899.804170] env[62569]: value = "task-1250343" [ 899.804170] env[62569]: _type = "Task" [ 899.804170] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.804473] env[62569]: DEBUG oslo_vmware.api [None req-f6f79371-bbdb-4d19-bd3b-66c4ed4687fe tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a18a03-791b-dbb5-0e5d-90ebe372436e, 'name': SearchDatastore_Task, 'duration_secs': 0.011555} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.804843] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f6f79371-bbdb-4d19-bd3b-66c4ed4687fe tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.808559] env[62569]: DEBUG nova.compute.manager [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 899.819934] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250343, 'name': CloneVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.954514] env[62569]: DEBUG nova.network.neutron [-] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.978367] env[62569]: DEBUG oslo_vmware.api [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250341, 'name': PowerOnVM_Task, 'duration_secs': 1.340935} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.978754] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 899.979065] env[62569]: INFO nova.compute.manager [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Took 10.29 seconds to spawn the instance on the hypervisor. [ 899.979204] env[62569]: DEBUG nova.compute.manager [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 899.980133] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f62a1dbd-c7cb-4d63-b877-760016de5266 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.089155] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "9c1c0379-3968-464b-b587-8fac704404d5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.089868] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "9c1c0379-3968-464b-b587-8fac704404d5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.089982] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "9c1c0379-3968-464b-b587-8fac704404d5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.090215] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "9c1c0379-3968-464b-b587-8fac704404d5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.090341] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "9c1c0379-3968-464b-b587-8fac704404d5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.092668] env[62569]: INFO nova.compute.manager [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Terminating instance [ 900.127182] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c01ec326-f5f9-4375-8c13-82313891cd34 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.138115] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e079c5ef-ff4c-4afa-8b3e-3f580f5840c8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.149804] env[62569]: ERROR nova.scheduler.client.report [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [req-0434b521-50ae-417d-a8d7-5092a4c57a95] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID fa06556a-5785-4014-b8bd-bc240a0cf716. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-0434b521-50ae-417d-a8d7-5092a4c57a95"}]} [ 900.178324] env[62569]: DEBUG nova.compute.manager [req-46ede12a-4307-43cc-8b50-bddd84edb065 req-c91ae0d2-69dc-4786-a6d4-85c428eb3df0 service nova] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Detach interface failed, port_id=19f8c74e-ebd3-4b14-9e7b-1539ce706834, reason: Instance 108d5bde-e463-4a69-816b-bdd3c03e13cd could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 900.179473] env[62569]: DEBUG nova.scheduler.client.report [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Refreshing inventories for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 900.195896] env[62569]: DEBUG nova.scheduler.client.report [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Updating ProviderTree inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 900.196167] env[62569]: DEBUG nova.compute.provider_tree [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 900.214835] env[62569]: DEBUG nova.scheduler.client.report [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Refreshing aggregate associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, aggregates: None {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 900.234386] env[62569]: DEBUG nova.scheduler.client.report [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Refreshing trait associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 900.286543] env[62569]: DEBUG oslo_vmware.api [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': task-1250342, 'name': PowerOffVM_Task, 'duration_secs': 0.308988} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.286897] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 900.287117] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 900.287416] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0dabfe66-1c42-45af-bbcd-e90743f7527c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.324580] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250343, 'name': CloneVM_Task} progress is 93%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.336495] env[62569]: DEBUG oslo_concurrency.lockutils [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.459679] env[62569]: INFO nova.compute.manager [-] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Took 1.65 seconds to deallocate network for instance. [ 900.502589] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24170b8d-f57f-447a-927b-f7a4dd8c0310 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.506460] env[62569]: INFO nova.compute.manager [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Took 20.10 seconds to build instance. [ 900.512670] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f29e186-c73d-4309-8e77-b8932c652207 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.549995] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-721d49e5-282c-4a26-8daf-7a4367d59b65 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.561086] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc03ecef-617d-4505-b7e3-2adcfa9a16db {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.580368] env[62569]: DEBUG nova.compute.provider_tree [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 900.597097] env[62569]: DEBUG nova.compute.manager [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 900.597351] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 900.598701] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b63b99f-59a0-485b-998c-5d43466f3df1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.614408] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 900.615488] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4f1683b3-cb0e-4fee-a72c-d6c8db4e90f5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.627936] env[62569]: DEBUG oslo_vmware.api [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 900.627936] env[62569]: value = "task-1250345" [ 900.627936] env[62569]: _type = "Task" [ 900.627936] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.642546] env[62569]: DEBUG oslo_vmware.api [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250345, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.644550] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 900.644859] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 900.645167] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Deleting the datastore file [datastore1] a6ee3547-3a3d-4480-b451-d1fddd829b39 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 900.645559] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8e9548e4-58a2-4bc3-a9c4-623d7f88572d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.655612] env[62569]: DEBUG oslo_vmware.api [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Waiting for the task: (returnval){ [ 900.655612] env[62569]: value = "task-1250346" [ 900.655612] env[62569]: _type = "Task" [ 900.655612] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.668688] env[62569]: DEBUG oslo_vmware.api [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': task-1250346, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.816928] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250343, 'name': CloneVM_Task} progress is 94%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.009437] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cab63e6c-0ad0-4154-b4c6-b38493d2d429 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.613s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.029496] env[62569]: INFO nova.compute.manager [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Took 0.57 seconds to detach 1 volumes for instance. [ 901.118765] env[62569]: ERROR nova.scheduler.client.report [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [req-149ec9fe-9b71-46c4-b460-953063b73b38] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID fa06556a-5785-4014-b8bd-bc240a0cf716. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-149ec9fe-9b71-46c4-b460-953063b73b38"}]} [ 901.139110] env[62569]: DEBUG oslo_vmware.api [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250345, 'name': PowerOffVM_Task, 'duration_secs': 0.398234} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.139315] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 901.139465] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 901.139998] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2123ec3a-93e8-4ea5-8cda-8eb067afeaa9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.142251] env[62569]: DEBUG nova.scheduler.client.report [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Refreshing inventories for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 901.161182] env[62569]: DEBUG nova.scheduler.client.report [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Updating ProviderTree inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 901.161801] env[62569]: DEBUG nova.compute.provider_tree [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 901.171303] env[62569]: DEBUG oslo_vmware.api [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Task: {'id': task-1250346, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.325049} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.171303] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 901.171513] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 901.171513] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 901.171909] env[62569]: INFO nova.compute.manager [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Took 1.45 seconds to destroy the instance on the hypervisor. [ 901.171978] env[62569]: DEBUG oslo.service.loopingcall [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.172361] env[62569]: DEBUG nova.compute.manager [-] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 901.172955] env[62569]: DEBUG nova.network.neutron [-] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 901.175315] env[62569]: DEBUG nova.scheduler.client.report [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Refreshing aggregate associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, aggregates: None {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 901.201515] env[62569]: DEBUG nova.scheduler.client.report [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Refreshing trait associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 901.225266] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 901.225266] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 901.225516] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Deleting the datastore file [datastore1] 9c1c0379-3968-464b-b587-8fac704404d5 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 901.225882] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5ce3eb78-587b-42e1-9c00-550d919ea052 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.234224] env[62569]: DEBUG oslo_vmware.api [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 901.234224] env[62569]: value = "task-1250348" [ 901.234224] env[62569]: _type = "Task" [ 901.234224] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.246905] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "6ccb78f3-d7f9-4090-9641-4633904efab8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.247324] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "6ccb78f3-d7f9-4090-9641-4633904efab8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.248859] env[62569]: DEBUG oslo_vmware.api [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250348, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.317402] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250343, 'name': CloneVM_Task} progress is 94%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.329611] env[62569]: DEBUG oslo_concurrency.lockutils [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Acquiring lock "f751dfdb-e950-4821-b70d-4db78e03b9b7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.329611] env[62569]: DEBUG oslo_concurrency.lockutils [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Lock "f751dfdb-e950-4821-b70d-4db78e03b9b7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.495475] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf17a5a-4df7-41b9-b430-685811a347d9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.504579] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3314853c-1bf5-4734-baee-c2ddaa96b0d1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.539376] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.540595] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e52ebba-7865-4bfe-8920-9710d40d5416 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.550570] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c96cbb5c-02dc-4342-acee-7aae68142a84 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.565407] env[62569]: DEBUG nova.compute.provider_tree [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 901.578008] env[62569]: DEBUG oslo_vmware.rw_handles [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523457f5-4c39-ea8d-cb0a-93de5f052b50/disk-0.vmdk. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 901.579296] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb947ff-9dfe-496f-aaa3-64a6487db44b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.586253] env[62569]: DEBUG oslo_vmware.rw_handles [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523457f5-4c39-ea8d-cb0a-93de5f052b50/disk-0.vmdk is in state: ready. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 901.586502] env[62569]: ERROR oslo_vmware.rw_handles [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523457f5-4c39-ea8d-cb0a-93de5f052b50/disk-0.vmdk due to incomplete transfer. [ 901.587017] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f7c54226-3724-4923-887b-272958303c7c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.595554] env[62569]: DEBUG oslo_vmware.rw_handles [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523457f5-4c39-ea8d-cb0a-93de5f052b50/disk-0.vmdk. {{(pid=62569) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 901.596574] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Uploaded image a618f103-681b-4bbf-a4d2-f503940d1cf8 to the Glance image server {{(pid=62569) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 901.598187] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Destroying the VM {{(pid=62569) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 901.599258] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-936a556d-73e8-43e3-9e94-1f8c770848ea {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.607050] env[62569]: DEBUG oslo_vmware.api [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 901.607050] env[62569]: value = "task-1250349" [ 901.607050] env[62569]: _type = "Task" [ 901.607050] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.617364] env[62569]: DEBUG oslo_vmware.api [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250349, 'name': Destroy_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.659953] env[62569]: DEBUG nova.compute.manager [req-b830399f-89d4-4595-b262-49c5307f8358 req-3030a94e-9d4a-408c-8432-7963101b7b65 service nova] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Received event network-vif-deleted-057050a7-da70-4b2f-bd56-b5de3a3dbb31 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 901.660224] env[62569]: INFO nova.compute.manager [req-b830399f-89d4-4595-b262-49c5307f8358 req-3030a94e-9d4a-408c-8432-7963101b7b65 service nova] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Neutron deleted interface 057050a7-da70-4b2f-bd56-b5de3a3dbb31; detaching it from the instance and deleting it from the info cache [ 901.660367] env[62569]: DEBUG nova.network.neutron [req-b830399f-89d4-4595-b262-49c5307f8358 req-3030a94e-9d4a-408c-8432-7963101b7b65 service nova] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.747124] env[62569]: DEBUG oslo_vmware.api [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250348, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.311654} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.747407] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 901.747590] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 901.748218] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 901.748218] env[62569]: INFO nova.compute.manager [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Took 1.15 seconds to destroy the instance on the hypervisor. [ 901.748218] env[62569]: DEBUG oslo.service.loopingcall [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.748396] env[62569]: DEBUG nova.compute.manager [-] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 901.749341] env[62569]: DEBUG nova.network.neutron [-] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 901.753728] env[62569]: DEBUG nova.compute.manager [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 901.821187] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250343, 'name': CloneVM_Task, 'duration_secs': 1.871931} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.821639] env[62569]: INFO nova.virt.vmwareapi.vmops [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Created linked-clone VM from snapshot [ 901.822789] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cb6bd44-4bd8-4078-9cc1-75be2fd9a75e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.833896] env[62569]: DEBUG nova.compute.manager [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 901.843563] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Uploading image 87099cdd-38b2-4615-944d-a91a20479717 {{(pid=62569) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 901.879413] env[62569]: DEBUG oslo_vmware.rw_handles [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 901.879413] env[62569]: value = "vm-269467" [ 901.879413] env[62569]: _type = "VirtualMachine" [ 901.879413] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 901.879413] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-893e48ec-b220-4e83-8ca5-334469e5a7c8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.889663] env[62569]: DEBUG oslo_vmware.rw_handles [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lease: (returnval){ [ 901.889663] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52efe05c-abac-4708-3baf-215e7660b0bd" [ 901.889663] env[62569]: _type = "HttpNfcLease" [ 901.889663] env[62569]: } obtained for exporting VM: (result){ [ 901.889663] env[62569]: value = "vm-269467" [ 901.889663] env[62569]: _type = "VirtualMachine" [ 901.889663] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 901.889962] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the lease: (returnval){ [ 901.889962] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52efe05c-abac-4708-3baf-215e7660b0bd" [ 901.889962] env[62569]: _type = "HttpNfcLease" [ 901.889962] env[62569]: } to be ready. {{(pid=62569) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 901.900173] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 901.900173] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52efe05c-abac-4708-3baf-215e7660b0bd" [ 901.900173] env[62569]: _type = "HttpNfcLease" [ 901.900173] env[62569]: } is initializing. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 901.973110] env[62569]: DEBUG nova.network.neutron [-] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.106555] env[62569]: DEBUG nova.scheduler.client.report [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Updated inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with generation 107 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 902.106914] env[62569]: DEBUG nova.compute.provider_tree [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Updating resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 generation from 107 to 108 during operation: update_inventory {{(pid=62569) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 902.107160] env[62569]: DEBUG nova.compute.provider_tree [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 902.122125] env[62569]: DEBUG oslo_vmware.api [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250349, 'name': Destroy_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.162869] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b04aa952-d019-4f6d-9ca7-484d70f562c4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.174268] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbea527f-9800-4373-b1c9-5f5933241d93 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.206946] env[62569]: DEBUG nova.compute.manager [req-b830399f-89d4-4595-b262-49c5307f8358 req-3030a94e-9d4a-408c-8432-7963101b7b65 service nova] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Detach interface failed, port_id=057050a7-da70-4b2f-bd56-b5de3a3dbb31, reason: Instance a6ee3547-3a3d-4480-b451-d1fddd829b39 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 902.255439] env[62569]: DEBUG nova.compute.manager [req-6e737e21-3741-4c19-843d-0dc53a3f6e21 req-b667d5cf-8433-4759-891a-89bb27481b30 service nova] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Received event network-changed-d64908ff-b5fa-42fe-8a49-1f39a27ca4b7 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 902.255654] env[62569]: DEBUG nova.compute.manager [req-6e737e21-3741-4c19-843d-0dc53a3f6e21 req-b667d5cf-8433-4759-891a-89bb27481b30 service nova] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Refreshing instance network info cache due to event network-changed-d64908ff-b5fa-42fe-8a49-1f39a27ca4b7. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 902.255878] env[62569]: DEBUG oslo_concurrency.lockutils [req-6e737e21-3741-4c19-843d-0dc53a3f6e21 req-b667d5cf-8433-4759-891a-89bb27481b30 service nova] Acquiring lock "refresh_cache-bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.260079] env[62569]: DEBUG oslo_concurrency.lockutils [req-6e737e21-3741-4c19-843d-0dc53a3f6e21 req-b667d5cf-8433-4759-891a-89bb27481b30 service nova] Acquired lock "refresh_cache-bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.260373] env[62569]: DEBUG nova.network.neutron [req-6e737e21-3741-4c19-843d-0dc53a3f6e21 req-b667d5cf-8433-4759-891a-89bb27481b30 service nova] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Refreshing network info cache for port d64908ff-b5fa-42fe-8a49-1f39a27ca4b7 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 902.285270] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.366298] env[62569]: DEBUG oslo_concurrency.lockutils [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.399981] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 902.399981] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52efe05c-abac-4708-3baf-215e7660b0bd" [ 902.399981] env[62569]: _type = "HttpNfcLease" [ 902.399981] env[62569]: } is ready. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 902.400363] env[62569]: DEBUG oslo_vmware.rw_handles [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 902.400363] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52efe05c-abac-4708-3baf-215e7660b0bd" [ 902.400363] env[62569]: _type = "HttpNfcLease" [ 902.400363] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 902.401244] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f770b347-9c1e-4f01-8288-6b778afcfa88 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.409349] env[62569]: DEBUG oslo_vmware.rw_handles [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f00fbb-6b48-cacf-6104-552d692aebfd/disk-0.vmdk from lease info. {{(pid=62569) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 902.409529] env[62569]: DEBUG oslo_vmware.rw_handles [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f00fbb-6b48-cacf-6104-552d692aebfd/disk-0.vmdk for reading. {{(pid=62569) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 902.476064] env[62569]: INFO nova.compute.manager [-] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Took 1.30 seconds to deallocate network for instance. [ 902.535794] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6658e147-9edb-4283-93ea-20eb2a74bab8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.576284] env[62569]: DEBUG nova.network.neutron [-] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.612611] env[62569]: DEBUG oslo_concurrency.lockutils [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 4.820s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.618756] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f6f79371-bbdb-4d19-bd3b-66c4ed4687fe tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 2.814s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.626826] env[62569]: DEBUG oslo_vmware.api [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250349, 'name': Destroy_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.634035] env[62569]: INFO nova.scheduler.client.report [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Deleted allocations for instance a0b406c3-9466-41bd-9de1-e675cab2ceef [ 902.983818] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.016931] env[62569]: DEBUG nova.network.neutron [req-6e737e21-3741-4c19-843d-0dc53a3f6e21 req-b667d5cf-8433-4759-891a-89bb27481b30 service nova] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Updated VIF entry in instance network info cache for port d64908ff-b5fa-42fe-8a49-1f39a27ca4b7. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 903.017858] env[62569]: DEBUG nova.network.neutron [req-6e737e21-3741-4c19-843d-0dc53a3f6e21 req-b667d5cf-8433-4759-891a-89bb27481b30 service nova] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Updating instance_info_cache with network_info: [{"id": "d64908ff-b5fa-42fe-8a49-1f39a27ca4b7", "address": "fa:16:3e:66:5c:fb", "network": {"id": "334a2e22-0249-48e3-a2f5-ac37df9771d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1048653043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f7fb3e698324f8e851aefbee74ac669", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f880ac2e-d532-4f54-87bb-998a8d1bca78", "external-id": "nsx-vlan-transportzone-491", "segmentation_id": 491, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd64908ff-b5", "ovs_interfaceid": "d64908ff-b5fa-42fe-8a49-1f39a27ca4b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.079070] env[62569]: INFO nova.compute.manager [-] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Took 1.33 seconds to deallocate network for instance. [ 903.127275] env[62569]: DEBUG oslo_vmware.api [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250349, 'name': Destroy_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.143212] env[62569]: DEBUG oslo_concurrency.lockutils [None req-570fce30-beb5-417c-9613-6b60a6eb534f tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "a0b406c3-9466-41bd-9de1-e675cab2ceef" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.684s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.408150] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c23ce48-88c0-4a33-878c-ad5c49a89f2f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.417544] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f81504-fca4-4c33-913c-b6a2b3f11143 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.459147] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ad2f055-330a-491f-b26f-da234786403a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.469926] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8125fa2d-3af8-4433-acea-c28600e3473e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.487644] env[62569]: DEBUG nova.compute.provider_tree [None req-f6f79371-bbdb-4d19-bd3b-66c4ed4687fe tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.520213] env[62569]: DEBUG oslo_concurrency.lockutils [req-6e737e21-3741-4c19-843d-0dc53a3f6e21 req-b667d5cf-8433-4759-891a-89bb27481b30 service nova] Releasing lock "refresh_cache-bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.587061] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.624132] env[62569]: DEBUG oslo_vmware.api [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250349, 'name': Destroy_Task, 'duration_secs': 1.544147} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.624503] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Destroyed the VM [ 903.624806] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Deleting Snapshot of the VM instance {{(pid=62569) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 903.625140] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4a4d2acd-4942-4ba2-8331-b8f59ad7f90c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.634560] env[62569]: DEBUG oslo_vmware.api [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 903.634560] env[62569]: value = "task-1250351" [ 903.634560] env[62569]: _type = "Task" [ 903.634560] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.648662] env[62569]: DEBUG oslo_vmware.api [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250351, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.691071] env[62569]: DEBUG nova.compute.manager [req-5e5d79dc-91ab-40bb-881b-70a16976c6f4 req-aee12f20-47ad-4002-a595-77093f666d22 service nova] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Received event network-vif-deleted-6ca336c0-76d8-4913-91e7-f7c65a3aa5c1 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 903.991506] env[62569]: DEBUG nova.scheduler.client.report [None req-f6f79371-bbdb-4d19-bd3b-66c4ed4687fe tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 904.148097] env[62569]: DEBUG oslo_vmware.api [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250351, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.650227] env[62569]: DEBUG oslo_vmware.api [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250351, 'name': RemoveSnapshot_Task, 'duration_secs': 0.547183} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.650778] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Deleted Snapshot of the VM instance {{(pid=62569) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 904.651150] env[62569]: INFO nova.compute.manager [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Took 22.23 seconds to snapshot the instance on the hypervisor. [ 905.003825] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f6f79371-bbdb-4d19-bd3b-66c4ed4687fe tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.385s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.006852] env[62569]: DEBUG oslo_concurrency.lockutils [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.670s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.008517] env[62569]: INFO nova.compute.claims [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 905.212732] env[62569]: DEBUG nova.compute.manager [None req-23c053e9-a541-4299-b1ed-55003495d4e4 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Found 2 images (rotation: 2) {{(pid=62569) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4923}} [ 905.397982] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.398312] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.573197] env[62569]: INFO nova.scheduler.client.report [None req-f6f79371-bbdb-4d19-bd3b-66c4ed4687fe tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Deleted allocation for migration 6322afb2-1484-48c6-8019-ecd2841d0d0b [ 905.901954] env[62569]: DEBUG nova.compute.manager [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 906.080673] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f6f79371-bbdb-4d19-bd3b-66c4ed4687fe tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "31ce29fa-4f60-4404-b830-21ad196f78b5" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.374s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.274393] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-084d648c-8cb1-4e73-b5d1-1d6e7c4bbae5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.283466] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d312e2d6-ddf9-45bd-87f1-c37b6dbc3991 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.321947] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3044e421-f04c-4775-b288-18498db5ff00 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.332266] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a8a4515-9437-4716-a8c4-c254755d66c3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.348376] env[62569]: DEBUG nova.compute.provider_tree [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 906.420434] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.647740] env[62569]: DEBUG nova.compute.manager [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 906.648643] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3c54d07-1735-4495-b753-edcd221be869 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.851211] env[62569]: DEBUG nova.scheduler.client.report [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 907.099804] env[62569]: DEBUG oslo_concurrency.lockutils [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "31ce29fa-4f60-4404-b830-21ad196f78b5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.100148] env[62569]: DEBUG oslo_concurrency.lockutils [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "31ce29fa-4f60-4404-b830-21ad196f78b5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.100409] env[62569]: DEBUG oslo_concurrency.lockutils [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "31ce29fa-4f60-4404-b830-21ad196f78b5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.100642] env[62569]: DEBUG oslo_concurrency.lockutils [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "31ce29fa-4f60-4404-b830-21ad196f78b5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.100799] env[62569]: DEBUG oslo_concurrency.lockutils [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "31ce29fa-4f60-4404-b830-21ad196f78b5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.103324] env[62569]: INFO nova.compute.manager [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Terminating instance [ 907.160080] env[62569]: INFO nova.compute.manager [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] instance snapshotting [ 907.160775] env[62569]: DEBUG nova.objects.instance [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lazy-loading 'flavor' on Instance uuid b47cd2d7-0cd2-41af-8ed1-a6dfca323516 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.356825] env[62569]: DEBUG oslo_concurrency.lockutils [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.350s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.357432] env[62569]: DEBUG nova.compute.manager [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 907.360756] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.822s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.361015] env[62569]: DEBUG nova.objects.instance [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lazy-loading 'resources' on Instance uuid 108d5bde-e463-4a69-816b-bdd3c03e13cd {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.525991] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "09eeb937-aabc-418d-9fd3-3c7a72f4ef75" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.526304] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "09eeb937-aabc-418d-9fd3-3c7a72f4ef75" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.606914] env[62569]: DEBUG nova.compute.manager [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 907.607395] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 907.608528] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d41c087-94f3-4dc2-971c-6ea260e3caa6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.617379] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 907.617676] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1233d4e8-5281-44c4-ba35-37125450a06e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.626122] env[62569]: DEBUG oslo_vmware.api [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 907.626122] env[62569]: value = "task-1250352" [ 907.626122] env[62569]: _type = "Task" [ 907.626122] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.636342] env[62569]: DEBUG oslo_vmware.api [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250352, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.667499] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6e2d709-20e7-4c13-8f8b-a96e89da830f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.689689] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-791ab64e-ce93-4748-8e48-6e790d44e9c0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.864482] env[62569]: DEBUG nova.compute.utils [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 907.870112] env[62569]: DEBUG nova.compute.manager [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 907.870112] env[62569]: DEBUG nova.network.neutron [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 907.910255] env[62569]: DEBUG nova.policy [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ab6d20862c54432cbafdda33ccc974ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '530ca7e7924743ab91a362a064a3111b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 908.028661] env[62569]: DEBUG nova.compute.manager [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 908.110507] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7651c3d4-9b1d-4eed-8660-a0cd8d578fd3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.119674] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8268afbf-b1a3-49c2-8d88-de3ec5f8e0c3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.155907] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-410e43f2-5ae6-4186-aaa2-11c9c3902bb3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.165399] env[62569]: DEBUG oslo_vmware.api [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250352, 'name': PowerOffVM_Task, 'duration_secs': 0.204557} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.167803] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 908.168042] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 908.168364] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-09fb6243-5c63-4f5d-8954-87e5ff18fa12 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.170904] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86a26de9-82c3-482d-8dd7-e05d77637982 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.185693] env[62569]: DEBUG nova.compute.provider_tree [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 908.202671] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Creating Snapshot of the VM instance {{(pid=62569) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 908.203046] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-4fdbf7d9-4aed-4340-a9b3-49797a511995 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.213140] env[62569]: DEBUG oslo_vmware.api [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 908.213140] env[62569]: value = "task-1250354" [ 908.213140] env[62569]: _type = "Task" [ 908.213140] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.225808] env[62569]: DEBUG oslo_vmware.api [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250354, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.246026] env[62569]: DEBUG nova.network.neutron [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Successfully created port: 91986023-125e-442b-b105-102630ec6036 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 908.249793] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 908.250033] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 908.250346] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Deleting the datastore file [datastore1] 31ce29fa-4f60-4404-b830-21ad196f78b5 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 908.250732] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5f7d47bb-7db5-4416-8456-8f59755264c8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.259778] env[62569]: DEBUG oslo_vmware.api [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 908.259778] env[62569]: value = "task-1250355" [ 908.259778] env[62569]: _type = "Task" [ 908.259778] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.269704] env[62569]: DEBUG oslo_vmware.api [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250355, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.372293] env[62569]: DEBUG nova.compute.manager [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 908.551110] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.692360] env[62569]: DEBUG nova.scheduler.client.report [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 908.724835] env[62569]: DEBUG oslo_vmware.api [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250354, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.773544] env[62569]: DEBUG oslo_vmware.api [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250355, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.258869} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.773823] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 908.774029] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 908.774219] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 908.774399] env[62569]: INFO nova.compute.manager [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Took 1.17 seconds to destroy the instance on the hypervisor. [ 908.774647] env[62569]: DEBUG oslo.service.loopingcall [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 908.774850] env[62569]: DEBUG nova.compute.manager [-] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 908.774945] env[62569]: DEBUG nova.network.neutron [-] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 909.079286] env[62569]: DEBUG nova.compute.manager [req-4adb77d0-aba3-4d38-ad1f-6099d8672288 req-377754b7-cbfb-4dcd-b749-6b5d57255ede service nova] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Received event network-vif-deleted-8e55dca6-f777-4333-9b49-feb5527d9a5e {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 909.079557] env[62569]: INFO nova.compute.manager [req-4adb77d0-aba3-4d38-ad1f-6099d8672288 req-377754b7-cbfb-4dcd-b749-6b5d57255ede service nova] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Neutron deleted interface 8e55dca6-f777-4333-9b49-feb5527d9a5e; detaching it from the instance and deleting it from the info cache [ 909.079557] env[62569]: DEBUG nova.network.neutron [req-4adb77d0-aba3-4d38-ad1f-6099d8672288 req-377754b7-cbfb-4dcd-b749-6b5d57255ede service nova] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.197356] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.836s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.200067] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.915s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.201348] env[62569]: INFO nova.compute.claims [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 909.223188] env[62569]: INFO nova.scheduler.client.report [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Deleted allocations for instance 108d5bde-e463-4a69-816b-bdd3c03e13cd [ 909.233420] env[62569]: DEBUG oslo_vmware.api [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250354, 'name': CreateSnapshot_Task, 'duration_secs': 0.67074} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.233600] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Created Snapshot of the VM instance {{(pid=62569) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 909.234563] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eecf97b-3626-4762-a9cb-59d9ba5af945 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.381645] env[62569]: DEBUG nova.compute.manager [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 909.407990] env[62569]: DEBUG nova.virt.hardware [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 909.408319] env[62569]: DEBUG nova.virt.hardware [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 909.408526] env[62569]: DEBUG nova.virt.hardware [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 909.408690] env[62569]: DEBUG nova.virt.hardware [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 909.408829] env[62569]: DEBUG nova.virt.hardware [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 909.408987] env[62569]: DEBUG nova.virt.hardware [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 909.409214] env[62569]: DEBUG nova.virt.hardware [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 909.409386] env[62569]: DEBUG nova.virt.hardware [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 909.409558] env[62569]: DEBUG nova.virt.hardware [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 909.409723] env[62569]: DEBUG nova.virt.hardware [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 909.409898] env[62569]: DEBUG nova.virt.hardware [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 909.410803] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64bfefa3-7664-4b0d-84ea-6e7cc6f0736d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.420106] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cdcf47c-6246-467b-9029-4cea18797c0b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.550052] env[62569]: DEBUG nova.network.neutron [-] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.583128] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-854f1f49-2578-4cfe-8a26-21a1f599a94b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.594899] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d596d58-5ade-40b1-99b2-358c63c4737b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.629287] env[62569]: DEBUG nova.compute.manager [req-4adb77d0-aba3-4d38-ad1f-6099d8672288 req-377754b7-cbfb-4dcd-b749-6b5d57255ede service nova] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Detach interface failed, port_id=8e55dca6-f777-4333-9b49-feb5527d9a5e, reason: Instance 31ce29fa-4f60-4404-b830-21ad196f78b5 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 909.732127] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5adaa44d-cde8-4c79-98a3-37713b245276 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "108d5bde-e463-4a69-816b-bdd3c03e13cd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.798s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.769367] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Creating linked-clone VM from snapshot {{(pid=62569) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 909.770397] env[62569]: DEBUG nova.network.neutron [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Successfully updated port: 91986023-125e-442b-b105-102630ec6036 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 909.772052] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1f062ee9-4401-4c85-9032-718c660167fe {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.782699] env[62569]: DEBUG oslo_vmware.api [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 909.782699] env[62569]: value = "task-1250356" [ 909.782699] env[62569]: _type = "Task" [ 909.782699] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.792382] env[62569]: DEBUG oslo_vmware.api [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250356, 'name': CloneVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.054944] env[62569]: INFO nova.compute.manager [-] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Took 1.28 seconds to deallocate network for instance. [ 910.278048] env[62569]: DEBUG oslo_concurrency.lockutils [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.278048] env[62569]: DEBUG oslo_concurrency.lockutils [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.278048] env[62569]: DEBUG nova.network.neutron [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 910.294922] env[62569]: DEBUG oslo_vmware.api [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250356, 'name': CloneVM_Task} progress is 94%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.450316] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-207582d9-b68d-42a8-a0de-07bade881445 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.462483] env[62569]: DEBUG oslo_vmware.rw_handles [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f00fbb-6b48-cacf-6104-552d692aebfd/disk-0.vmdk. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 910.463423] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbec9d4e-974e-4e96-8b6f-d82a3a74cf69 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.467315] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22a2bded-b0b7-4b82-aad5-67f72e871645 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.473900] env[62569]: DEBUG oslo_vmware.rw_handles [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f00fbb-6b48-cacf-6104-552d692aebfd/disk-0.vmdk is in state: ready. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 910.474864] env[62569]: ERROR oslo_vmware.rw_handles [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f00fbb-6b48-cacf-6104-552d692aebfd/disk-0.vmdk due to incomplete transfer. [ 910.474864] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-0d69b744-f391-4175-891f-0b6f09efba79 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.502009] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de92b770-17be-419a-bbe2-563016be6bae {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.511724] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ecdfa57-9cd4-44bf-a16e-3c382ede31ef {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.516216] env[62569]: DEBUG oslo_vmware.rw_handles [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f00fbb-6b48-cacf-6104-552d692aebfd/disk-0.vmdk. {{(pid=62569) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 910.516389] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Uploaded image 87099cdd-38b2-4615-944d-a91a20479717 to the Glance image server {{(pid=62569) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 910.519302] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Destroying the VM {{(pid=62569) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 910.519567] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-04349334-a837-4895-853b-188a6734865d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.531506] env[62569]: DEBUG nova.compute.provider_tree [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 910.534252] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 910.534252] env[62569]: value = "task-1250357" [ 910.534252] env[62569]: _type = "Task" [ 910.534252] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.544268] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250357, 'name': Destroy_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.562410] env[62569]: DEBUG oslo_concurrency.lockutils [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.795993] env[62569]: DEBUG oslo_vmware.api [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250356, 'name': CloneVM_Task} progress is 95%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.814369] env[62569]: DEBUG nova.network.neutron [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 910.937357] env[62569]: DEBUG nova.network.neutron [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Updating instance_info_cache with network_info: [{"id": "91986023-125e-442b-b105-102630ec6036", "address": "fa:16:3e:18:83:9b", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91986023-12", "ovs_interfaceid": "91986023-125e-442b-b105-102630ec6036", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.036488] env[62569]: DEBUG nova.scheduler.client.report [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 911.048033] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "af26f440-3515-4be7-9a03-8a0404c152d9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.048342] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "af26f440-3515-4be7-9a03-8a0404c152d9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.053356] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250357, 'name': Destroy_Task, 'duration_secs': 0.362151} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.054117] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Destroyed the VM [ 911.054370] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Deleting Snapshot of the VM instance {{(pid=62569) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 911.054845] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-954fd1ce-91e1-476a-9c89-e60982ee94f3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.062364] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 911.062364] env[62569]: value = "task-1250358" [ 911.062364] env[62569]: _type = "Task" [ 911.062364] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.071375] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250358, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.107817] env[62569]: DEBUG nova.compute.manager [req-ffe8e1d7-275c-480d-a2f3-946b859d4d90 req-1edd669e-9633-4364-892b-41b6def3803e service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Received event network-vif-plugged-91986023-125e-442b-b105-102630ec6036 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 911.108470] env[62569]: DEBUG oslo_concurrency.lockutils [req-ffe8e1d7-275c-480d-a2f3-946b859d4d90 req-1edd669e-9633-4364-892b-41b6def3803e service nova] Acquiring lock "188cac03-4034-4a02-973a-fb1906399fd1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.108741] env[62569]: DEBUG oslo_concurrency.lockutils [req-ffe8e1d7-275c-480d-a2f3-946b859d4d90 req-1edd669e-9633-4364-892b-41b6def3803e service nova] Lock "188cac03-4034-4a02-973a-fb1906399fd1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.108968] env[62569]: DEBUG oslo_concurrency.lockutils [req-ffe8e1d7-275c-480d-a2f3-946b859d4d90 req-1edd669e-9633-4364-892b-41b6def3803e service nova] Lock "188cac03-4034-4a02-973a-fb1906399fd1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.109218] env[62569]: DEBUG nova.compute.manager [req-ffe8e1d7-275c-480d-a2f3-946b859d4d90 req-1edd669e-9633-4364-892b-41b6def3803e service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] No waiting events found dispatching network-vif-plugged-91986023-125e-442b-b105-102630ec6036 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 911.109430] env[62569]: WARNING nova.compute.manager [req-ffe8e1d7-275c-480d-a2f3-946b859d4d90 req-1edd669e-9633-4364-892b-41b6def3803e service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Received unexpected event network-vif-plugged-91986023-125e-442b-b105-102630ec6036 for instance with vm_state building and task_state spawning. [ 911.109639] env[62569]: DEBUG nova.compute.manager [req-ffe8e1d7-275c-480d-a2f3-946b859d4d90 req-1edd669e-9633-4364-892b-41b6def3803e service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Received event network-changed-91986023-125e-442b-b105-102630ec6036 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 911.109841] env[62569]: DEBUG nova.compute.manager [req-ffe8e1d7-275c-480d-a2f3-946b859d4d90 req-1edd669e-9633-4364-892b-41b6def3803e service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Refreshing instance network info cache due to event network-changed-91986023-125e-442b-b105-102630ec6036. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 911.110060] env[62569]: DEBUG oslo_concurrency.lockutils [req-ffe8e1d7-275c-480d-a2f3-946b859d4d90 req-1edd669e-9633-4364-892b-41b6def3803e service nova] Acquiring lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.294930] env[62569]: DEBUG oslo_vmware.api [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250356, 'name': CloneVM_Task, 'duration_secs': 1.339261} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.295262] env[62569]: INFO nova.virt.vmwareapi.vmops [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Created linked-clone VM from snapshot [ 911.296050] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42a5b102-15d5-4fc2-a103-3b53db355d16 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.303585] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Uploading image d01a021f-098f-43be-b983-f770c15e1c56 {{(pid=62569) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 911.329330] env[62569]: DEBUG oslo_vmware.rw_handles [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 911.329330] env[62569]: value = "vm-269469" [ 911.329330] env[62569]: _type = "VirtualMachine" [ 911.329330] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 911.329613] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8d1aef16-1165-4e69-bfd1-d5872285ff35 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.337671] env[62569]: DEBUG oslo_vmware.rw_handles [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lease: (returnval){ [ 911.337671] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5244f369-bbf9-73bd-241d-57a1512d1c17" [ 911.337671] env[62569]: _type = "HttpNfcLease" [ 911.337671] env[62569]: } obtained for exporting VM: (result){ [ 911.337671] env[62569]: value = "vm-269469" [ 911.337671] env[62569]: _type = "VirtualMachine" [ 911.337671] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 911.337959] env[62569]: DEBUG oslo_vmware.api [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the lease: (returnval){ [ 911.337959] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5244f369-bbf9-73bd-241d-57a1512d1c17" [ 911.337959] env[62569]: _type = "HttpNfcLease" [ 911.337959] env[62569]: } to be ready. {{(pid=62569) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 911.344425] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 911.344425] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5244f369-bbf9-73bd-241d-57a1512d1c17" [ 911.344425] env[62569]: _type = "HttpNfcLease" [ 911.344425] env[62569]: } is initializing. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 911.440767] env[62569]: DEBUG oslo_concurrency.lockutils [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.441212] env[62569]: DEBUG nova.compute.manager [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Instance network_info: |[{"id": "91986023-125e-442b-b105-102630ec6036", "address": "fa:16:3e:18:83:9b", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91986023-12", "ovs_interfaceid": "91986023-125e-442b-b105-102630ec6036", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 911.441566] env[62569]: DEBUG oslo_concurrency.lockutils [req-ffe8e1d7-275c-480d-a2f3-946b859d4d90 req-1edd669e-9633-4364-892b-41b6def3803e service nova] Acquired lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.441761] env[62569]: DEBUG nova.network.neutron [req-ffe8e1d7-275c-480d-a2f3-946b859d4d90 req-1edd669e-9633-4364-892b-41b6def3803e service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Refreshing network info cache for port 91986023-125e-442b-b105-102630ec6036 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 911.444078] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:18:83:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '099fe970-c61f-4480-bed4-ae4f485fd82a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '91986023-125e-442b-b105-102630ec6036', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 911.452182] env[62569]: DEBUG oslo.service.loopingcall [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.453177] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 911.453502] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-819af3a5-7b80-437f-868a-7fbe25fb38e5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.474738] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 911.474738] env[62569]: value = "task-1250360" [ 911.474738] env[62569]: _type = "Task" [ 911.474738] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.482945] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250360, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.545416] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.346s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.545957] env[62569]: DEBUG nova.compute.manager [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 911.548616] env[62569]: DEBUG oslo_concurrency.lockutils [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.182s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.550073] env[62569]: INFO nova.compute.claims [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 911.553137] env[62569]: DEBUG nova.compute.manager [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 911.573257] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250358, 'name': RemoveSnapshot_Task, 'duration_secs': 0.408261} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.573553] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Deleted Snapshot of the VM instance {{(pid=62569) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 911.573849] env[62569]: DEBUG nova.compute.manager [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 911.574852] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02da29e2-9b0d-4428-880d-fe3cd121b19e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.846705] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 911.846705] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5244f369-bbf9-73bd-241d-57a1512d1c17" [ 911.846705] env[62569]: _type = "HttpNfcLease" [ 911.846705] env[62569]: } is ready. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 911.847054] env[62569]: DEBUG oslo_vmware.rw_handles [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 911.847054] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5244f369-bbf9-73bd-241d-57a1512d1c17" [ 911.847054] env[62569]: _type = "HttpNfcLease" [ 911.847054] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 911.847801] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b812660-280e-436d-bae8-35acc477eaa7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.856552] env[62569]: DEBUG oslo_vmware.rw_handles [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529df833-a3a7-236b-a90a-8e6455159856/disk-0.vmdk from lease info. {{(pid=62569) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 911.856552] env[62569]: DEBUG oslo_vmware.rw_handles [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529df833-a3a7-236b-a90a-8e6455159856/disk-0.vmdk for reading. {{(pid=62569) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 911.948413] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-237e7da3-d1a7-48e9-a9b9-ebb9fec42994 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.991348] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250360, 'name': CreateVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.054485] env[62569]: DEBUG nova.compute.utils [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 912.055983] env[62569]: DEBUG nova.compute.manager [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 912.056169] env[62569]: DEBUG nova.network.neutron [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 912.092308] env[62569]: INFO nova.compute.manager [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Shelve offloading [ 912.118549] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.146871] env[62569]: DEBUG nova.policy [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6dd052494f244fd78050c5d26f497720', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2c6a4814c234d53bbd4844cd482fbae', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 912.270915] env[62569]: DEBUG nova.network.neutron [req-ffe8e1d7-275c-480d-a2f3-946b859d4d90 req-1edd669e-9633-4364-892b-41b6def3803e service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Updated VIF entry in instance network info cache for port 91986023-125e-442b-b105-102630ec6036. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 912.271611] env[62569]: DEBUG nova.network.neutron [req-ffe8e1d7-275c-480d-a2f3-946b859d4d90 req-1edd669e-9633-4364-892b-41b6def3803e service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Updating instance_info_cache with network_info: [{"id": "91986023-125e-442b-b105-102630ec6036", "address": "fa:16:3e:18:83:9b", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91986023-12", "ovs_interfaceid": "91986023-125e-442b-b105-102630ec6036", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.483972] env[62569]: DEBUG nova.network.neutron [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Successfully created port: 37c1cdc3-9e78-4f21-b75e-219d6166d47b {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 912.490246] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250360, 'name': CreateVM_Task, 'duration_secs': 0.579138} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.490480] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 912.491422] env[62569]: DEBUG oslo_concurrency.lockutils [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.491678] env[62569]: DEBUG oslo_concurrency.lockutils [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.492103] env[62569]: DEBUG oslo_concurrency.lockutils [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 912.492945] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-335757d4-e29e-4a33-9332-2da88da1890d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.500090] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 912.500090] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]521e18fc-e090-a609-96d5-dbcc3aed53c4" [ 912.500090] env[62569]: _type = "Task" [ 912.500090] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.510442] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]521e18fc-e090-a609-96d5-dbcc3aed53c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.560918] env[62569]: DEBUG nova.compute.manager [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 912.598041] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 912.598625] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9c1b2e26-c8cc-4c54-9d47-8c5c00668b60 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.609100] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 912.609100] env[62569]: value = "task-1250361" [ 912.609100] env[62569]: _type = "Task" [ 912.609100] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.621843] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] VM already powered off {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 912.622105] env[62569]: DEBUG nova.compute.manager [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 912.623133] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa64e01f-0a7c-4912-9e71-f81632bb08cf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.634796] env[62569]: DEBUG oslo_concurrency.lockutils [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.634988] env[62569]: DEBUG oslo_concurrency.lockutils [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquired lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.635208] env[62569]: DEBUG nova.network.neutron [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 912.774570] env[62569]: DEBUG oslo_concurrency.lockutils [req-ffe8e1d7-275c-480d-a2f3-946b859d4d90 req-1edd669e-9633-4364-892b-41b6def3803e service nova] Releasing lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.842283] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e577c78-5280-4014-bb54-e2801fde9c21 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.850328] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-564e60c3-d5a6-4c52-b9b4-b6fee84311a0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.884636] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-695708f6-3d13-42af-81d4-2546e378f141 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.894046] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-903f1744-e92e-4a38-8d5f-7f6da20c0198 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.915202] env[62569]: DEBUG nova.compute.provider_tree [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.010740] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]521e18fc-e090-a609-96d5-dbcc3aed53c4, 'name': SearchDatastore_Task, 'duration_secs': 0.016114} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.011462] env[62569]: DEBUG oslo_concurrency.lockutils [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.012135] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 913.012135] env[62569]: DEBUG oslo_concurrency.lockutils [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.012291] env[62569]: DEBUG oslo_concurrency.lockutils [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.012403] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 913.012686] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-30252e5a-ee9d-4b72-8bb1-6975e9828c4a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.024189] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 913.024552] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 913.025317] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a5820f8-1905-4f63-9fd6-0ad1f5668f87 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.031613] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 913.031613] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a3a24c-6987-ac6c-601b-d94c16994cd9" [ 913.031613] env[62569]: _type = "Task" [ 913.031613] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.040927] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a3a24c-6987-ac6c-601b-d94c16994cd9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.145405] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de27057a-ab19-4aa8-a0b6-bbc243731817 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.146235] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de27057a-ab19-4aa8-a0b6-bbc243731817 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.146416] env[62569]: INFO nova.compute.manager [None req-de27057a-ab19-4aa8-a0b6-bbc243731817 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Rebooting instance [ 913.421021] env[62569]: DEBUG nova.scheduler.client.report [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 913.423627] env[62569]: DEBUG nova.network.neutron [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Updating instance_info_cache with network_info: [{"id": "cfddbae7-9ab0-4bc6-aea1-46a0cde743fc", "address": "fa:16:3e:b7:d6:80", "network": {"id": "406dd658-8168-4e73-9b97-861929bfaa2e", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-693938985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de8605118d744a93bce54e897aa849b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfddbae7-9a", "ovs_interfaceid": "cfddbae7-9ab0-4bc6-aea1-46a0cde743fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.542857] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a3a24c-6987-ac6c-601b-d94c16994cd9, 'name': SearchDatastore_Task, 'duration_secs': 0.012501} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.544143] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-346e3cbe-9bc7-4f3b-85d2-9e214a7331d8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.550853] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 913.550853] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]526f4f93-464e-1af9-7d11-30a0b69b67f8" [ 913.550853] env[62569]: _type = "Task" [ 913.550853] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.560093] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]526f4f93-464e-1af9-7d11-30a0b69b67f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.572350] env[62569]: DEBUG nova.compute.manager [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 913.597503] env[62569]: DEBUG nova.virt.hardware [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 913.597909] env[62569]: DEBUG nova.virt.hardware [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 913.598112] env[62569]: DEBUG nova.virt.hardware [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 913.598349] env[62569]: DEBUG nova.virt.hardware [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 913.598513] env[62569]: DEBUG nova.virt.hardware [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 913.599100] env[62569]: DEBUG nova.virt.hardware [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 913.599211] env[62569]: DEBUG nova.virt.hardware [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 913.599466] env[62569]: DEBUG nova.virt.hardware [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 913.599771] env[62569]: DEBUG nova.virt.hardware [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 913.600015] env[62569]: DEBUG nova.virt.hardware [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 913.600255] env[62569]: DEBUG nova.virt.hardware [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 913.601247] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9261f845-d884-4eaa-9f39-f49ccecc96de {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.610430] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a1831fa-d9d3-4fb2-91c5-3fe3e7183da1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.664638] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de27057a-ab19-4aa8-a0b6-bbc243731817 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.664896] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de27057a-ab19-4aa8-a0b6-bbc243731817 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.665106] env[62569]: DEBUG nova.network.neutron [None req-de27057a-ab19-4aa8-a0b6-bbc243731817 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 913.926689] env[62569]: DEBUG oslo_concurrency.lockutils [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.378s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.927274] env[62569]: DEBUG nova.compute.manager [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 913.934661] env[62569]: DEBUG oslo_concurrency.lockutils [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Releasing lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.937176] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.953s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.937470] env[62569]: DEBUG nova.objects.instance [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Lazy-loading 'resources' on Instance uuid a6ee3547-3a3d-4480-b451-d1fddd829b39 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 913.977396] env[62569]: DEBUG nova.compute.manager [req-55c06d8f-e134-49a2-9c07-7b65310bb7c8 req-3bf4bd70-f663-4158-aac9-c74e00f6a712 service nova] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Received event network-vif-plugged-37c1cdc3-9e78-4f21-b75e-219d6166d47b {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 913.977396] env[62569]: DEBUG oslo_concurrency.lockutils [req-55c06d8f-e134-49a2-9c07-7b65310bb7c8 req-3bf4bd70-f663-4158-aac9-c74e00f6a712 service nova] Acquiring lock "6ccb78f3-d7f9-4090-9641-4633904efab8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.977396] env[62569]: DEBUG oslo_concurrency.lockutils [req-55c06d8f-e134-49a2-9c07-7b65310bb7c8 req-3bf4bd70-f663-4158-aac9-c74e00f6a712 service nova] Lock "6ccb78f3-d7f9-4090-9641-4633904efab8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.977599] env[62569]: DEBUG oslo_concurrency.lockutils [req-55c06d8f-e134-49a2-9c07-7b65310bb7c8 req-3bf4bd70-f663-4158-aac9-c74e00f6a712 service nova] Lock "6ccb78f3-d7f9-4090-9641-4633904efab8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.977681] env[62569]: DEBUG nova.compute.manager [req-55c06d8f-e134-49a2-9c07-7b65310bb7c8 req-3bf4bd70-f663-4158-aac9-c74e00f6a712 service nova] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] No waiting events found dispatching network-vif-plugged-37c1cdc3-9e78-4f21-b75e-219d6166d47b {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 913.977861] env[62569]: WARNING nova.compute.manager [req-55c06d8f-e134-49a2-9c07-7b65310bb7c8 req-3bf4bd70-f663-4158-aac9-c74e00f6a712 service nova] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Received unexpected event network-vif-plugged-37c1cdc3-9e78-4f21-b75e-219d6166d47b for instance with vm_state building and task_state spawning. [ 914.067252] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]526f4f93-464e-1af9-7d11-30a0b69b67f8, 'name': SearchDatastore_Task, 'duration_secs': 0.018391} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.067787] env[62569]: DEBUG oslo_concurrency.lockutils [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.068175] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 188cac03-4034-4a02-973a-fb1906399fd1/188cac03-4034-4a02-973a-fb1906399fd1.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 914.068564] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7debd7fb-cf80-4b7a-81c0-e1471914d6f6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.076718] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 914.076718] env[62569]: value = "task-1250362" [ 914.076718] env[62569]: _type = "Task" [ 914.076718] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.086619] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250362, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.129257] env[62569]: DEBUG nova.network.neutron [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Successfully updated port: 37c1cdc3-9e78-4f21-b75e-219d6166d47b {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 914.399892] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 914.404709] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63efa9e4-44cf-467d-bfdd-14e3400a51e4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.410127] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 914.410445] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-976a2c9f-eb58-4fd2-9bee-507835e8971b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.436077] env[62569]: DEBUG nova.compute.utils [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 914.440108] env[62569]: DEBUG nova.compute.manager [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 914.440108] env[62569]: DEBUG nova.network.neutron [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 914.505160] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 914.505420] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 914.505604] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Deleting the datastore file [datastore2] b6e5eefc-8c06-445b-a3af-9404578b6179 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 914.505894] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-05f5a4d4-3972-40a8-a8f8-d71a1aad598b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.515891] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 914.515891] env[62569]: value = "task-1250364" [ 914.515891] env[62569]: _type = "Task" [ 914.515891] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.521838] env[62569]: DEBUG nova.policy [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3b6252b56f1b4509a173cd4d04e2cfbe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2bd3b44104ae4eeabde752ccf8fd020c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 914.533491] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250364, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.587709] env[62569]: DEBUG nova.network.neutron [None req-de27057a-ab19-4aa8-a0b6-bbc243731817 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Updating instance_info_cache with network_info: [{"id": "39e71a4f-3b81-4b34-8cc2-2a5352eca4ce", "address": "fa:16:3e:4b:e2:a8", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap39e71a4f-3b", "ovs_interfaceid": "39e71a4f-3b81-4b34-8cc2-2a5352eca4ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.597744] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250362, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.634800] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "refresh_cache-6ccb78f3-d7f9-4090-9641-4633904efab8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.634955] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired lock "refresh_cache-6ccb78f3-d7f9-4090-9641-4633904efab8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.635122] env[62569]: DEBUG nova.network.neutron [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 914.730547] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33f0fc06-0b41-4839-8776-71f800e75744 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.739170] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-421e5c5f-fa33-4c5e-a45d-ea158779eeb1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.778609] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40d2692a-6f5b-428c-b00a-c15f3c4e6d84 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.787479] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-258f0ceb-9062-4012-b119-ea0d15a3e107 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.803091] env[62569]: DEBUG nova.compute.provider_tree [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.837352] env[62569]: DEBUG nova.network.neutron [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Successfully created port: 89885c41-7de0-49a3-8606-d500eb1308f3 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 914.943066] env[62569]: DEBUG nova.compute.manager [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 915.028118] env[62569]: DEBUG oslo_vmware.api [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250364, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.309435} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.028438] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 915.028705] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 915.028939] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 915.051879] env[62569]: INFO nova.scheduler.client.report [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Deleted allocations for instance b6e5eefc-8c06-445b-a3af-9404578b6179 [ 915.089692] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250362, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.554823} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.089986] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 188cac03-4034-4a02-973a-fb1906399fd1/188cac03-4034-4a02-973a-fb1906399fd1.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 915.090502] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 915.091281] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de27057a-ab19-4aa8-a0b6-bbc243731817 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.092481] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ff98f424-d906-4491-a537-7baf0081d2c2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.103559] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 915.103559] env[62569]: value = "task-1250365" [ 915.103559] env[62569]: _type = "Task" [ 915.103559] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.113947] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250365, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.174639] env[62569]: DEBUG nova.network.neutron [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 915.306066] env[62569]: DEBUG nova.scheduler.client.report [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 915.319653] env[62569]: DEBUG nova.network.neutron [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Updating instance_info_cache with network_info: [{"id": "37c1cdc3-9e78-4f21-b75e-219d6166d47b", "address": "fa:16:3e:db:d8:2f", "network": {"id": "884d3248-07f1-4cf7-8ceb-b94b34622384", "bridge": "br-int", "label": "tempest-ServersTestJSON-1882023778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2c6a4814c234d53bbd4844cd482fbae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37c1cdc3-9e", "ovs_interfaceid": "37c1cdc3-9e78-4f21-b75e-219d6166d47b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.557939] env[62569]: DEBUG oslo_concurrency.lockutils [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.599642] env[62569]: DEBUG nova.compute.manager [None req-de27057a-ab19-4aa8-a0b6-bbc243731817 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 915.600290] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27444094-a24c-4086-b0d5-e4a1b0640346 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.616759] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250365, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.812331] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.875s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.814898] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.228s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.815222] env[62569]: DEBUG nova.objects.instance [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lazy-loading 'resources' on Instance uuid 9c1c0379-3968-464b-b587-8fac704404d5 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 915.822879] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Releasing lock "refresh_cache-6ccb78f3-d7f9-4090-9641-4633904efab8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.822993] env[62569]: DEBUG nova.compute.manager [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Instance network_info: |[{"id": "37c1cdc3-9e78-4f21-b75e-219d6166d47b", "address": "fa:16:3e:db:d8:2f", "network": {"id": "884d3248-07f1-4cf7-8ceb-b94b34622384", "bridge": "br-int", "label": "tempest-ServersTestJSON-1882023778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2c6a4814c234d53bbd4844cd482fbae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37c1cdc3-9e", "ovs_interfaceid": "37c1cdc3-9e78-4f21-b75e-219d6166d47b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 915.823419] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:db:d8:2f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e9fa4744-8702-4973-b911-ee18192a3e4b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '37c1cdc3-9e78-4f21-b75e-219d6166d47b', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 915.834293] env[62569]: DEBUG oslo.service.loopingcall [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.835566] env[62569]: INFO nova.scheduler.client.report [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Deleted allocations for instance a6ee3547-3a3d-4480-b451-d1fddd829b39 [ 915.837273] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 915.837577] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-edadb9c9-c5ea-40b1-bbde-8dedd07bbefe {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.862854] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 915.862854] env[62569]: value = "task-1250366" [ 915.862854] env[62569]: _type = "Task" [ 915.862854] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.872863] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250366, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.953854] env[62569]: DEBUG nova.compute.manager [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 915.980595] env[62569]: DEBUG nova.virt.hardware [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 915.980853] env[62569]: DEBUG nova.virt.hardware [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 915.981022] env[62569]: DEBUG nova.virt.hardware [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 915.981306] env[62569]: DEBUG nova.virt.hardware [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 915.981546] env[62569]: DEBUG nova.virt.hardware [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 915.981639] env[62569]: DEBUG nova.virt.hardware [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 915.981815] env[62569]: DEBUG nova.virt.hardware [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 915.981982] env[62569]: DEBUG nova.virt.hardware [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 915.982210] env[62569]: DEBUG nova.virt.hardware [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 915.982354] env[62569]: DEBUG nova.virt.hardware [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 915.982522] env[62569]: DEBUG nova.virt.hardware [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 915.983427] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-835118b5-03b3-4599-8f7d-bee59d470de7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.992404] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae77cad1-f306-4d4c-95db-73a4f0905c8e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.015950] env[62569]: DEBUG nova.compute.manager [req-1a9174dd-ec78-417c-be83-c0d381a8a489 req-c7f2173e-8d31-4706-bde5-5950a8a9b730 service nova] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Received event network-changed-37c1cdc3-9e78-4f21-b75e-219d6166d47b {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 916.016195] env[62569]: DEBUG nova.compute.manager [req-1a9174dd-ec78-417c-be83-c0d381a8a489 req-c7f2173e-8d31-4706-bde5-5950a8a9b730 service nova] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Refreshing instance network info cache due to event network-changed-37c1cdc3-9e78-4f21-b75e-219d6166d47b. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 916.016395] env[62569]: DEBUG oslo_concurrency.lockutils [req-1a9174dd-ec78-417c-be83-c0d381a8a489 req-c7f2173e-8d31-4706-bde5-5950a8a9b730 service nova] Acquiring lock "refresh_cache-6ccb78f3-d7f9-4090-9641-4633904efab8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.016545] env[62569]: DEBUG oslo_concurrency.lockutils [req-1a9174dd-ec78-417c-be83-c0d381a8a489 req-c7f2173e-8d31-4706-bde5-5950a8a9b730 service nova] Acquired lock "refresh_cache-6ccb78f3-d7f9-4090-9641-4633904efab8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.016711] env[62569]: DEBUG nova.network.neutron [req-1a9174dd-ec78-417c-be83-c0d381a8a489 req-c7f2173e-8d31-4706-bde5-5950a8a9b730 service nova] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Refreshing network info cache for port 37c1cdc3-9e78-4f21-b75e-219d6166d47b {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 916.117641] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250365, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.523709} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.117956] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 916.118764] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf22f44c-d0e1-4303-8fef-3dd7a3e73530 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.142486] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 188cac03-4034-4a02-973a-fb1906399fd1/188cac03-4034-4a02-973a-fb1906399fd1.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 916.142771] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-567905dc-820c-45d2-b15f-71f0e89d6b79 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.167256] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 916.167256] env[62569]: value = "task-1250367" [ 916.167256] env[62569]: _type = "Task" [ 916.167256] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.177022] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250367, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.359791] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d8e1939-d4a5-43c3-9d1e-55e43eb0edc1 tempest-ServerMetadataTestJSON-280218454 tempest-ServerMetadataTestJSON-280218454-project-member] Lock "a6ee3547-3a3d-4480-b451-d1fddd829b39" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.155s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.375835] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250366, 'name': CreateVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.398113] env[62569]: DEBUG nova.network.neutron [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Successfully updated port: 89885c41-7de0-49a3-8606-d500eb1308f3 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 916.549042] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46d28f3a-9632-45d4-abe0-cd48e53585f5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.556427] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-326cf214-2a9a-491b-b08f-a29f72dba1fa {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.585801] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7a1acee-6b72-4f50-9434-fd7545c10c63 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.597561] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9658fa6b-80d0-4068-94f8-edd1d56741ca {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.611711] env[62569]: DEBUG nova.compute.provider_tree [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.620033] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d0ee48b-88f7-4701-9e19-203ce91e9853 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.627012] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-de27057a-ab19-4aa8-a0b6-bbc243731817 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Doing hard reboot of VM {{(pid=62569) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 916.627546] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-6c1d300a-b1e0-4503-a969-1bea39d2682d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.634139] env[62569]: DEBUG oslo_vmware.api [None req-de27057a-ab19-4aa8-a0b6-bbc243731817 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 916.634139] env[62569]: value = "task-1250368" [ 916.634139] env[62569]: _type = "Task" [ 916.634139] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.645351] env[62569]: DEBUG oslo_vmware.api [None req-de27057a-ab19-4aa8-a0b6-bbc243731817 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250368, 'name': ResetVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.681267] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250367, 'name': ReconfigVM_Task, 'duration_secs': 0.296708} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.681879] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 188cac03-4034-4a02-973a-fb1906399fd1/188cac03-4034-4a02-973a-fb1906399fd1.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 916.682375] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-04625235-23c5-4ff9-9717-c9e60cfd831a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.689366] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 916.689366] env[62569]: value = "task-1250369" [ 916.689366] env[62569]: _type = "Task" [ 916.689366] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.697898] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250369, 'name': Rename_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.752492] env[62569]: DEBUG nova.network.neutron [req-1a9174dd-ec78-417c-be83-c0d381a8a489 req-c7f2173e-8d31-4706-bde5-5950a8a9b730 service nova] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Updated VIF entry in instance network info cache for port 37c1cdc3-9e78-4f21-b75e-219d6166d47b. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 916.752866] env[62569]: DEBUG nova.network.neutron [req-1a9174dd-ec78-417c-be83-c0d381a8a489 req-c7f2173e-8d31-4706-bde5-5950a8a9b730 service nova] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Updating instance_info_cache with network_info: [{"id": "37c1cdc3-9e78-4f21-b75e-219d6166d47b", "address": "fa:16:3e:db:d8:2f", "network": {"id": "884d3248-07f1-4cf7-8ceb-b94b34622384", "bridge": "br-int", "label": "tempest-ServersTestJSON-1882023778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2c6a4814c234d53bbd4844cd482fbae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37c1cdc3-9e", "ovs_interfaceid": "37c1cdc3-9e78-4f21-b75e-219d6166d47b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.834969] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "b6e5eefc-8c06-445b-a3af-9404578b6179" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.873755] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250366, 'name': CreateVM_Task, 'duration_secs': 0.532854} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.873957] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 916.874674] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.874894] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.875226] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 916.875511] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d7d0353-57e1-4055-ba0b-2e1ff82d7119 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.880079] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 916.880079] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5273c3f7-15ba-ce1b-86ae-9f8b251863eb" [ 916.880079] env[62569]: _type = "Task" [ 916.880079] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.887957] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5273c3f7-15ba-ce1b-86ae-9f8b251863eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.902139] env[62569]: DEBUG oslo_concurrency.lockutils [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Acquiring lock "refresh_cache-f751dfdb-e950-4821-b70d-4db78e03b9b7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.902262] env[62569]: DEBUG oslo_concurrency.lockutils [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Acquired lock "refresh_cache-f751dfdb-e950-4821-b70d-4db78e03b9b7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.902427] env[62569]: DEBUG nova.network.neutron [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 917.116913] env[62569]: DEBUG nova.scheduler.client.report [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 917.147328] env[62569]: DEBUG oslo_vmware.api [None req-de27057a-ab19-4aa8-a0b6-bbc243731817 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250368, 'name': ResetVM_Task, 'duration_secs': 0.093889} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.147328] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-de27057a-ab19-4aa8-a0b6-bbc243731817 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Did hard reboot of VM {{(pid=62569) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 917.147328] env[62569]: DEBUG nova.compute.manager [None req-de27057a-ab19-4aa8-a0b6-bbc243731817 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 917.147328] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65b748e8-c685-4568-bada-a38c351166a1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.202643] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250369, 'name': Rename_Task, 'duration_secs': 0.172865} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.202643] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 917.202643] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5b64434b-e0d7-4a1f-a01d-2358f00b63ed {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.210967] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 917.210967] env[62569]: value = "task-1250370" [ 917.210967] env[62569]: _type = "Task" [ 917.210967] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.219437] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250370, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.255393] env[62569]: DEBUG oslo_concurrency.lockutils [req-1a9174dd-ec78-417c-be83-c0d381a8a489 req-c7f2173e-8d31-4706-bde5-5950a8a9b730 service nova] Releasing lock "refresh_cache-6ccb78f3-d7f9-4090-9641-4633904efab8" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.255673] env[62569]: DEBUG nova.compute.manager [req-1a9174dd-ec78-417c-be83-c0d381a8a489 req-c7f2173e-8d31-4706-bde5-5950a8a9b730 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Received event network-vif-unplugged-cfddbae7-9ab0-4bc6-aea1-46a0cde743fc {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 917.256108] env[62569]: DEBUG oslo_concurrency.lockutils [req-1a9174dd-ec78-417c-be83-c0d381a8a489 req-c7f2173e-8d31-4706-bde5-5950a8a9b730 service nova] Acquiring lock "b6e5eefc-8c06-445b-a3af-9404578b6179-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.256108] env[62569]: DEBUG oslo_concurrency.lockutils [req-1a9174dd-ec78-417c-be83-c0d381a8a489 req-c7f2173e-8d31-4706-bde5-5950a8a9b730 service nova] Lock "b6e5eefc-8c06-445b-a3af-9404578b6179-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.256305] env[62569]: DEBUG oslo_concurrency.lockutils [req-1a9174dd-ec78-417c-be83-c0d381a8a489 req-c7f2173e-8d31-4706-bde5-5950a8a9b730 service nova] Lock "b6e5eefc-8c06-445b-a3af-9404578b6179-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.256486] env[62569]: DEBUG nova.compute.manager [req-1a9174dd-ec78-417c-be83-c0d381a8a489 req-c7f2173e-8d31-4706-bde5-5950a8a9b730 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] No waiting events found dispatching network-vif-unplugged-cfddbae7-9ab0-4bc6-aea1-46a0cde743fc {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 917.256664] env[62569]: WARNING nova.compute.manager [req-1a9174dd-ec78-417c-be83-c0d381a8a489 req-c7f2173e-8d31-4706-bde5-5950a8a9b730 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Received unexpected event network-vif-unplugged-cfddbae7-9ab0-4bc6-aea1-46a0cde743fc for instance with vm_state shelved_offloaded and task_state unshelving. [ 917.256833] env[62569]: DEBUG nova.compute.manager [req-1a9174dd-ec78-417c-be83-c0d381a8a489 req-c7f2173e-8d31-4706-bde5-5950a8a9b730 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Received event network-changed-cfddbae7-9ab0-4bc6-aea1-46a0cde743fc {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 917.256993] env[62569]: DEBUG nova.compute.manager [req-1a9174dd-ec78-417c-be83-c0d381a8a489 req-c7f2173e-8d31-4706-bde5-5950a8a9b730 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Refreshing instance network info cache due to event network-changed-cfddbae7-9ab0-4bc6-aea1-46a0cde743fc. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 917.257198] env[62569]: DEBUG oslo_concurrency.lockutils [req-1a9174dd-ec78-417c-be83-c0d381a8a489 req-c7f2173e-8d31-4706-bde5-5950a8a9b730 service nova] Acquiring lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.257341] env[62569]: DEBUG oslo_concurrency.lockutils [req-1a9174dd-ec78-417c-be83-c0d381a8a489 req-c7f2173e-8d31-4706-bde5-5950a8a9b730 service nova] Acquired lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.257500] env[62569]: DEBUG nova.network.neutron [req-1a9174dd-ec78-417c-be83-c0d381a8a489 req-c7f2173e-8d31-4706-bde5-5950a8a9b730 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Refreshing network info cache for port cfddbae7-9ab0-4bc6-aea1-46a0cde743fc {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 917.393571] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5273c3f7-15ba-ce1b-86ae-9f8b251863eb, 'name': SearchDatastore_Task, 'duration_secs': 0.010796} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.393925] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.394198] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 917.394465] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.394623] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.394829] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 917.395128] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-de838386-6f5d-4780-bbe2-59f72c970d66 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.407100] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 917.407369] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 917.408609] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5439218b-b958-4c14-bb9a-66ca2bf53832 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.416708] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 917.416708] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528c8d05-0b66-2ad7-0b82-1878376c6675" [ 917.416708] env[62569]: _type = "Task" [ 917.416708] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.428693] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528c8d05-0b66-2ad7-0b82-1878376c6675, 'name': SearchDatastore_Task, 'duration_secs': 0.009863} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.429605] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04ea8151-8a33-4259-b997-7f0755c982d1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.436184] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 917.436184] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5204a508-f987-1ab0-ae91-aaf15ab52ce4" [ 917.436184] env[62569]: _type = "Task" [ 917.436184] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.444409] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5204a508-f987-1ab0-ae91-aaf15ab52ce4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.447593] env[62569]: DEBUG nova.network.neutron [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 917.606648] env[62569]: DEBUG nova.network.neutron [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Updating instance_info_cache with network_info: [{"id": "89885c41-7de0-49a3-8606-d500eb1308f3", "address": "fa:16:3e:57:17:9f", "network": {"id": "db1d1d1c-186c-4f30-950c-2906fa2fc43a", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-660695237-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2bd3b44104ae4eeabde752ccf8fd020c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89885c41-7d", "ovs_interfaceid": "89885c41-7de0-49a3-8606-d500eb1308f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.622309] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.807s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.626358] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.206s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.628244] env[62569]: INFO nova.compute.claims [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 917.649328] env[62569]: INFO nova.scheduler.client.report [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Deleted allocations for instance 9c1c0379-3968-464b-b587-8fac704404d5 [ 917.661213] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de27057a-ab19-4aa8-a0b6-bbc243731817 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.514s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.723172] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250370, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.956319] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5204a508-f987-1ab0-ae91-aaf15ab52ce4, 'name': SearchDatastore_Task, 'duration_secs': 0.00991} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.956647] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.956934] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 6ccb78f3-d7f9-4090-9641-4633904efab8/6ccb78f3-d7f9-4090-9641-4633904efab8.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 917.957244] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f29e45dc-aea1-45ba-b25f-752a23cef296 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.964523] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 917.964523] env[62569]: value = "task-1250371" [ 917.964523] env[62569]: _type = "Task" [ 917.964523] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.973541] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250371, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.055958] env[62569]: DEBUG nova.compute.manager [req-ee9fb821-a71c-4aaf-ab8b-b82c269c9fa8 req-8daa4ff1-2288-47ea-ad1a-78f41a868419 service nova] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Received event network-vif-plugged-89885c41-7de0-49a3-8606-d500eb1308f3 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 918.056344] env[62569]: DEBUG oslo_concurrency.lockutils [req-ee9fb821-a71c-4aaf-ab8b-b82c269c9fa8 req-8daa4ff1-2288-47ea-ad1a-78f41a868419 service nova] Acquiring lock "f751dfdb-e950-4821-b70d-4db78e03b9b7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.058307] env[62569]: DEBUG oslo_concurrency.lockutils [req-ee9fb821-a71c-4aaf-ab8b-b82c269c9fa8 req-8daa4ff1-2288-47ea-ad1a-78f41a868419 service nova] Lock "f751dfdb-e950-4821-b70d-4db78e03b9b7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.058307] env[62569]: DEBUG oslo_concurrency.lockutils [req-ee9fb821-a71c-4aaf-ab8b-b82c269c9fa8 req-8daa4ff1-2288-47ea-ad1a-78f41a868419 service nova] Lock "f751dfdb-e950-4821-b70d-4db78e03b9b7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.002s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.058307] env[62569]: DEBUG nova.compute.manager [req-ee9fb821-a71c-4aaf-ab8b-b82c269c9fa8 req-8daa4ff1-2288-47ea-ad1a-78f41a868419 service nova] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] No waiting events found dispatching network-vif-plugged-89885c41-7de0-49a3-8606-d500eb1308f3 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 918.058527] env[62569]: WARNING nova.compute.manager [req-ee9fb821-a71c-4aaf-ab8b-b82c269c9fa8 req-8daa4ff1-2288-47ea-ad1a-78f41a868419 service nova] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Received unexpected event network-vif-plugged-89885c41-7de0-49a3-8606-d500eb1308f3 for instance with vm_state building and task_state spawning. [ 918.058731] env[62569]: DEBUG nova.compute.manager [req-ee9fb821-a71c-4aaf-ab8b-b82c269c9fa8 req-8daa4ff1-2288-47ea-ad1a-78f41a868419 service nova] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Received event network-changed-89885c41-7de0-49a3-8606-d500eb1308f3 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 918.058838] env[62569]: DEBUG nova.compute.manager [req-ee9fb821-a71c-4aaf-ab8b-b82c269c9fa8 req-8daa4ff1-2288-47ea-ad1a-78f41a868419 service nova] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Refreshing instance network info cache due to event network-changed-89885c41-7de0-49a3-8606-d500eb1308f3. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 918.059034] env[62569]: DEBUG oslo_concurrency.lockutils [req-ee9fb821-a71c-4aaf-ab8b-b82c269c9fa8 req-8daa4ff1-2288-47ea-ad1a-78f41a868419 service nova] Acquiring lock "refresh_cache-f751dfdb-e950-4821-b70d-4db78e03b9b7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.108807] env[62569]: DEBUG nova.network.neutron [req-1a9174dd-ec78-417c-be83-c0d381a8a489 req-c7f2173e-8d31-4706-bde5-5950a8a9b730 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Updated VIF entry in instance network info cache for port cfddbae7-9ab0-4bc6-aea1-46a0cde743fc. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 918.108807] env[62569]: DEBUG nova.network.neutron [req-1a9174dd-ec78-417c-be83-c0d381a8a489 req-c7f2173e-8d31-4706-bde5-5950a8a9b730 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Updating instance_info_cache with network_info: [{"id": "cfddbae7-9ab0-4bc6-aea1-46a0cde743fc", "address": "fa:16:3e:b7:d6:80", "network": {"id": "406dd658-8168-4e73-9b97-861929bfaa2e", "bridge": null, "label": "tempest-ServersNegativeTestJSON-693938985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de8605118d744a93bce54e897aa849b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapcfddbae7-9a", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.111886] env[62569]: DEBUG oslo_concurrency.lockutils [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Releasing lock "refresh_cache-f751dfdb-e950-4821-b70d-4db78e03b9b7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.114388] env[62569]: DEBUG nova.compute.manager [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Instance network_info: |[{"id": "89885c41-7de0-49a3-8606-d500eb1308f3", "address": "fa:16:3e:57:17:9f", "network": {"id": "db1d1d1c-186c-4f30-950c-2906fa2fc43a", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-660695237-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2bd3b44104ae4eeabde752ccf8fd020c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89885c41-7d", "ovs_interfaceid": "89885c41-7de0-49a3-8606-d500eb1308f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 918.114388] env[62569]: DEBUG oslo_concurrency.lockutils [req-ee9fb821-a71c-4aaf-ab8b-b82c269c9fa8 req-8daa4ff1-2288-47ea-ad1a-78f41a868419 service nova] Acquired lock "refresh_cache-f751dfdb-e950-4821-b70d-4db78e03b9b7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.114388] env[62569]: DEBUG nova.network.neutron [req-ee9fb821-a71c-4aaf-ab8b-b82c269c9fa8 req-8daa4ff1-2288-47ea-ad1a-78f41a868419 service nova] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Refreshing network info cache for port 89885c41-7de0-49a3-8606-d500eb1308f3 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 918.114921] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:57:17:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5514c5a3-1294-40ad-ae96-29d5c24a3d95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '89885c41-7de0-49a3-8606-d500eb1308f3', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 918.131154] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Creating folder: Project (2bd3b44104ae4eeabde752ccf8fd020c). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 918.136358] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c6fc9ef4-0751-4021-841d-664abaaf2e3f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.152576] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Created folder: Project (2bd3b44104ae4eeabde752ccf8fd020c) in parent group-v269330. [ 918.153588] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Creating folder: Instances. Parent ref: group-v269472. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 918.154246] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d5221194-1ff3-488e-a539-f90e7684cd62 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.160806] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0e765521-3d0c-46ad-9ddd-19431d3b06b5 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "9c1c0379-3968-464b-b587-8fac704404d5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.070s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.169029] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Created folder: Instances in parent group-v269472. [ 918.169029] env[62569]: DEBUG oslo.service.loopingcall [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 918.169029] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 918.169029] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-284e525f-b3f0-4b05-8579-d6ffebc692a3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.195847] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 918.195847] env[62569]: value = "task-1250374" [ 918.195847] env[62569]: _type = "Task" [ 918.195847] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.204922] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250374, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.222155] env[62569]: DEBUG oslo_vmware.api [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250370, 'name': PowerOnVM_Task, 'duration_secs': 0.627368} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.222470] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 918.222715] env[62569]: INFO nova.compute.manager [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Took 8.84 seconds to spawn the instance on the hypervisor. [ 918.222988] env[62569]: DEBUG nova.compute.manager [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 918.223851] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85c8d9dd-965d-4140-ab7c-6fd3a22e9b39 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.484587] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250371, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.545682] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "05db90a6-3faf-4878-b782-ed17df47ed5f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.545682] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "05db90a6-3faf-4878-b782-ed17df47ed5f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.545682] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "05db90a6-3faf-4878-b782-ed17df47ed5f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.545682] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "05db90a6-3faf-4878-b782-ed17df47ed5f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.545909] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "05db90a6-3faf-4878-b782-ed17df47ed5f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.548654] env[62569]: INFO nova.compute.manager [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Terminating instance [ 918.613807] env[62569]: DEBUG oslo_concurrency.lockutils [req-1a9174dd-ec78-417c-be83-c0d381a8a489 req-c7f2173e-8d31-4706-bde5-5950a8a9b730 service nova] Releasing lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.706936] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250374, 'name': CreateVM_Task, 'duration_secs': 0.402618} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.707248] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 918.710663] env[62569]: DEBUG oslo_concurrency.lockutils [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.710862] env[62569]: DEBUG oslo_concurrency.lockutils [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.711382] env[62569]: DEBUG oslo_concurrency.lockutils [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 918.712291] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2b5d51c-8fea-4088-b563-d3c200fa0a37 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.720024] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Waiting for the task: (returnval){ [ 918.720024] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529ae8c3-0bf2-b259-307f-c9b5b512fd10" [ 918.720024] env[62569]: _type = "Task" [ 918.720024] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.733480] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529ae8c3-0bf2-b259-307f-c9b5b512fd10, 'name': SearchDatastore_Task, 'duration_secs': 0.011885} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.733480] env[62569]: DEBUG oslo_concurrency.lockutils [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.734159] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 918.735462] env[62569]: DEBUG oslo_concurrency.lockutils [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.737018] env[62569]: DEBUG oslo_concurrency.lockutils [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.737018] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 918.742777] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a6650f66-a2ea-4b89-b557-060521055ea5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.751572] env[62569]: DEBUG nova.network.neutron [req-ee9fb821-a71c-4aaf-ab8b-b82c269c9fa8 req-8daa4ff1-2288-47ea-ad1a-78f41a868419 service nova] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Updated VIF entry in instance network info cache for port 89885c41-7de0-49a3-8606-d500eb1308f3. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 918.751928] env[62569]: DEBUG nova.network.neutron [req-ee9fb821-a71c-4aaf-ab8b-b82c269c9fa8 req-8daa4ff1-2288-47ea-ad1a-78f41a868419 service nova] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Updating instance_info_cache with network_info: [{"id": "89885c41-7de0-49a3-8606-d500eb1308f3", "address": "fa:16:3e:57:17:9f", "network": {"id": "db1d1d1c-186c-4f30-950c-2906fa2fc43a", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-660695237-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2bd3b44104ae4eeabde752ccf8fd020c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89885c41-7d", "ovs_interfaceid": "89885c41-7de0-49a3-8606-d500eb1308f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.753529] env[62569]: INFO nova.compute.manager [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Took 18.44 seconds to build instance. [ 918.765048] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 918.765048] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 918.765333] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-196c13e2-1817-4ef9-b0c4-50dca3e089dc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.777025] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Waiting for the task: (returnval){ [ 918.777025] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528b26e7-f402-c7cd-2884-d8a9727071da" [ 918.777025] env[62569]: _type = "Task" [ 918.777025] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.787296] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528b26e7-f402-c7cd-2884-d8a9727071da, 'name': SearchDatastore_Task, 'duration_secs': 0.011011} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.791074] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f30167b9-31c8-4fc7-b252-d516f5a837ed {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.797314] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Waiting for the task: (returnval){ [ 918.797314] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5274da1c-8370-e4af-3682-9e64c28d5832" [ 918.797314] env[62569]: _type = "Task" [ 918.797314] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.810415] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5274da1c-8370-e4af-3682-9e64c28d5832, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.980300] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250371, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.641079} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.980464] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 6ccb78f3-d7f9-4090-9641-4633904efab8/6ccb78f3-d7f9-4090-9641-4633904efab8.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 918.980679] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 918.983609] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a6b88324-9067-4fe4-8633-83fc2d7d974e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.990655] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 918.990655] env[62569]: value = "task-1250375" [ 918.990655] env[62569]: _type = "Task" [ 918.990655] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.001517] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250375, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.013620] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c7c4d3d-98e6-4f42-a1e3-52fd34c7019a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.021577] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a0f2686-2cba-4031-9c3f-ba752f35c607 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.057781] env[62569]: DEBUG nova.compute.manager [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 919.058039] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 919.059087] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16da162d-e4c1-4d59-8366-2ad3e537081f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.063922] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab2fa60-ea67-47ee-81f8-a281d0d47cfb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.069364] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 919.071706] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4ed4e468-a08e-4b04-a6ab-053983cdc384 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.074470] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-463a1f2b-878f-4076-83d0-ae5cd7e205e6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.088234] env[62569]: DEBUG nova.compute.provider_tree [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 919.090975] env[62569]: DEBUG oslo_vmware.api [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 919.090975] env[62569]: value = "task-1250376" [ 919.090975] env[62569]: _type = "Task" [ 919.090975] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.099065] env[62569]: DEBUG oslo_vmware.api [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250376, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.200665] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "d1fdff70-0401-49ca-bbbb-ef5eda266c9c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.200665] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "d1fdff70-0401-49ca-bbbb-ef5eda266c9c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.255744] env[62569]: DEBUG oslo_concurrency.lockutils [req-ee9fb821-a71c-4aaf-ab8b-b82c269c9fa8 req-8daa4ff1-2288-47ea-ad1a-78f41a868419 service nova] Releasing lock "refresh_cache-f751dfdb-e950-4821-b70d-4db78e03b9b7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.256443] env[62569]: DEBUG oslo_concurrency.lockutils [None req-920a1971-63ce-4b70-9918-85cd7cbfafc3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "188cac03-4034-4a02-973a-fb1906399fd1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.957s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.307817] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5274da1c-8370-e4af-3682-9e64c28d5832, 'name': SearchDatastore_Task, 'duration_secs': 0.012019} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.308133] env[62569]: DEBUG oslo_concurrency.lockutils [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.308465] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] f751dfdb-e950-4821-b70d-4db78e03b9b7/f751dfdb-e950-4821-b70d-4db78e03b9b7.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 919.308751] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5f354689-556c-46c7-afc4-759ba85048f6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.316500] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Waiting for the task: (returnval){ [ 919.316500] env[62569]: value = "task-1250377" [ 919.316500] env[62569]: _type = "Task" [ 919.316500] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.325563] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Task: {'id': task-1250377, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.500106] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250375, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084934} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.500254] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 919.501152] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-267f8fcd-fd7f-44d8-8a39-a78b72ba94e9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.524419] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 6ccb78f3-d7f9-4090-9641-4633904efab8/6ccb78f3-d7f9-4090-9641-4633904efab8.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 919.524762] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d681ba1-d193-4ff6-9253-e82ba1cb526d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.545745] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 919.545745] env[62569]: value = "task-1250378" [ 919.545745] env[62569]: _type = "Task" [ 919.545745] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.555831] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250378, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.593633] env[62569]: DEBUG nova.scheduler.client.report [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 919.609440] env[62569]: DEBUG oslo_vmware.api [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250376, 'name': PowerOffVM_Task, 'duration_secs': 0.219754} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.610534] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 919.610710] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 919.611033] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d4ed17be-64aa-4592-a5ec-b4679b095712 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.677717] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 919.677980] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 919.678221] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Deleting the datastore file [datastore1] 05db90a6-3faf-4878-b782-ed17df47ed5f {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 919.678939] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-379612c3-a71b-4e85-923a-947a31b46d22 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.690133] env[62569]: DEBUG oslo_vmware.api [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 919.690133] env[62569]: value = "task-1250380" [ 919.690133] env[62569]: _type = "Task" [ 919.690133] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.700859] env[62569]: DEBUG oslo_vmware.api [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250380, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.703369] env[62569]: DEBUG nova.compute.manager [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 919.826682] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Task: {'id': task-1250377, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.060446] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250378, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.102912] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.476s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.103583] env[62569]: DEBUG nova.compute.manager [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 920.106893] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.556s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.108478] env[62569]: INFO nova.compute.claims [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 920.200413] env[62569]: DEBUG oslo_vmware.api [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250380, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.300274} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.200669] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 920.200855] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 920.201144] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 920.201806] env[62569]: INFO nova.compute.manager [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Took 1.14 seconds to destroy the instance on the hypervisor. [ 920.201806] env[62569]: DEBUG oslo.service.loopingcall [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 920.201806] env[62569]: DEBUG nova.compute.manager [-] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 920.201989] env[62569]: DEBUG nova.network.neutron [-] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 920.225445] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.311423] env[62569]: DEBUG nova.compute.manager [req-944a39fe-007f-450b-9d3e-90f352aff02c req-08513460-25a7-407a-b865-df15c860272b service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Received event network-changed-91986023-125e-442b-b105-102630ec6036 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 920.311689] env[62569]: DEBUG nova.compute.manager [req-944a39fe-007f-450b-9d3e-90f352aff02c req-08513460-25a7-407a-b865-df15c860272b service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Refreshing instance network info cache due to event network-changed-91986023-125e-442b-b105-102630ec6036. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 920.312045] env[62569]: DEBUG oslo_concurrency.lockutils [req-944a39fe-007f-450b-9d3e-90f352aff02c req-08513460-25a7-407a-b865-df15c860272b service nova] Acquiring lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.312277] env[62569]: DEBUG oslo_concurrency.lockutils [req-944a39fe-007f-450b-9d3e-90f352aff02c req-08513460-25a7-407a-b865-df15c860272b service nova] Acquired lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.312511] env[62569]: DEBUG nova.network.neutron [req-944a39fe-007f-450b-9d3e-90f352aff02c req-08513460-25a7-407a-b865-df15c860272b service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Refreshing network info cache for port 91986023-125e-442b-b105-102630ec6036 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 920.332764] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Task: {'id': task-1250377, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.526732} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.332764] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] f751dfdb-e950-4821-b70d-4db78e03b9b7/f751dfdb-e950-4821-b70d-4db78e03b9b7.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 920.332987] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 920.333159] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e179e392-e917-417f-b10f-a5adbc661ce9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.340590] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Waiting for the task: (returnval){ [ 920.340590] env[62569]: value = "task-1250381" [ 920.340590] env[62569]: _type = "Task" [ 920.340590] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.350308] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Task: {'id': task-1250381, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.557851] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250378, 'name': ReconfigVM_Task, 'duration_secs': 0.575656} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.558211] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 6ccb78f3-d7f9-4090-9641-4633904efab8/6ccb78f3-d7f9-4090-9641-4633904efab8.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 920.558897] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-57363049-8131-4a33-a246-cc17047a0425 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.567148] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 920.567148] env[62569]: value = "task-1250382" [ 920.567148] env[62569]: _type = "Task" [ 920.567148] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.581855] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250382, 'name': Rename_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.584901] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.585229] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.613312] env[62569]: DEBUG oslo_vmware.rw_handles [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529df833-a3a7-236b-a90a-8e6455159856/disk-0.vmdk. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 920.614733] env[62569]: DEBUG nova.compute.utils [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 920.618368] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7a8884f-d79d-417a-a7ba-c66dc4b1c6a4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.621605] env[62569]: DEBUG nova.compute.manager [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 920.621829] env[62569]: DEBUG nova.network.neutron [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 920.631043] env[62569]: DEBUG oslo_vmware.rw_handles [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529df833-a3a7-236b-a90a-8e6455159856/disk-0.vmdk is in state: ready. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 920.631043] env[62569]: ERROR oslo_vmware.rw_handles [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529df833-a3a7-236b-a90a-8e6455159856/disk-0.vmdk due to incomplete transfer. [ 920.631043] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5df00cab-6687-40c5-a084-48649c8a4410 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.637168] env[62569]: DEBUG oslo_vmware.rw_handles [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529df833-a3a7-236b-a90a-8e6455159856/disk-0.vmdk. {{(pid=62569) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 920.637376] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Uploaded image d01a021f-098f-43be-b983-f770c15e1c56 to the Glance image server {{(pid=62569) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 920.639698] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Destroying the VM {{(pid=62569) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 920.639957] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-59963755-9388-45f2-aeeb-048d29ad8d23 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.645953] env[62569]: DEBUG oslo_vmware.api [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 920.645953] env[62569]: value = "task-1250383" [ 920.645953] env[62569]: _type = "Task" [ 920.645953] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.653745] env[62569]: DEBUG oslo_vmware.api [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250383, 'name': Destroy_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.666102] env[62569]: DEBUG nova.policy [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '75adf33d63ac4d07b358ed72daf8f12d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e9ea55b00b574a7db0530efcb498ca2b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 920.854164] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Task: {'id': task-1250381, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079978} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.854164] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 920.854164] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fffd02c6-d74d-4efe-a267-534efa51cc74 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.877960] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] f751dfdb-e950-4821-b70d-4db78e03b9b7/f751dfdb-e950-4821-b70d-4db78e03b9b7.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 920.880242] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8a5c055a-f906-42f0-8caf-5982d2119cab {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.900665] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Waiting for the task: (returnval){ [ 920.900665] env[62569]: value = "task-1250384" [ 920.900665] env[62569]: _type = "Task" [ 920.900665] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.908650] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Task: {'id': task-1250384, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.973548] env[62569]: DEBUG nova.network.neutron [-] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.034781] env[62569]: DEBUG nova.network.neutron [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Successfully created port: f9665162-0615-45cd-b3c2-32be38c9af60 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 921.081154] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250382, 'name': Rename_Task, 'duration_secs': 0.149886} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.081456] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 921.081715] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f442bee3-d3d5-418e-a307-9100da04a413 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.087703] env[62569]: DEBUG nova.compute.manager [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 921.092888] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 921.092888] env[62569]: value = "task-1250385" [ 921.092888] env[62569]: _type = "Task" [ 921.092888] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.101382] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250385, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.122855] env[62569]: DEBUG nova.compute.manager [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 921.155994] env[62569]: DEBUG oslo_vmware.api [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250383, 'name': Destroy_Task, 'duration_secs': 0.342246} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.159266] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Destroyed the VM [ 921.159266] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Deleting Snapshot of the VM instance {{(pid=62569) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 921.159266] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-649ddf60-94fd-4a75-8519-62f13c41dc5f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.172034] env[62569]: DEBUG oslo_vmware.api [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 921.172034] env[62569]: value = "task-1250386" [ 921.172034] env[62569]: _type = "Task" [ 921.172034] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.179860] env[62569]: DEBUG oslo_vmware.api [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250386, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.232402] env[62569]: DEBUG nova.network.neutron [req-944a39fe-007f-450b-9d3e-90f352aff02c req-08513460-25a7-407a-b865-df15c860272b service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Updated VIF entry in instance network info cache for port 91986023-125e-442b-b105-102630ec6036. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 921.232847] env[62569]: DEBUG nova.network.neutron [req-944a39fe-007f-450b-9d3e-90f352aff02c req-08513460-25a7-407a-b865-df15c860272b service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Updating instance_info_cache with network_info: [{"id": "91986023-125e-442b-b105-102630ec6036", "address": "fa:16:3e:18:83:9b", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91986023-12", "ovs_interfaceid": "91986023-125e-442b-b105-102630ec6036", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.413623] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Task: {'id': task-1250384, 'name': ReconfigVM_Task, 'duration_secs': 0.332326} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.413826] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Reconfigured VM instance instance-00000059 to attach disk [datastore2] f751dfdb-e950-4821-b70d-4db78e03b9b7/f751dfdb-e950-4821-b70d-4db78e03b9b7.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 921.415325] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5631aab8-fdef-48de-8e6c-d41e71bd94c1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.417956] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c1d749e-6777-4b77-bacb-003230fbf661 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.427891] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-216c375f-0d81-44b7-ad1e-7223565147ab {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.431032] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Waiting for the task: (returnval){ [ 921.431032] env[62569]: value = "task-1250387" [ 921.431032] env[62569]: _type = "Task" [ 921.431032] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.464042] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60459a46-a03a-4a0f-9431-3ff08f80c0a2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.466687] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Task: {'id': task-1250387, 'name': Rename_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.472197] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62a58a2a-187b-4e4b-9e93-5675763c71ed {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.476443] env[62569]: INFO nova.compute.manager [-] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Took 1.27 seconds to deallocate network for instance. [ 921.488711] env[62569]: DEBUG nova.compute.provider_tree [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 921.605215] env[62569]: DEBUG oslo_vmware.api [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250385, 'name': PowerOnVM_Task, 'duration_secs': 0.472287} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.605500] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 921.605710] env[62569]: INFO nova.compute.manager [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Took 8.03 seconds to spawn the instance on the hypervisor. [ 921.605891] env[62569]: DEBUG nova.compute.manager [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 921.606687] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c671c8-220a-465b-890f-f9db020e13a8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.610204] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.682033] env[62569]: DEBUG oslo_vmware.api [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250386, 'name': RemoveSnapshot_Task} progress is 65%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.735557] env[62569]: DEBUG oslo_concurrency.lockutils [req-944a39fe-007f-450b-9d3e-90f352aff02c req-08513460-25a7-407a-b865-df15c860272b service nova] Releasing lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.940728] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Task: {'id': task-1250387, 'name': Rename_Task, 'duration_secs': 0.149679} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.940966] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 921.941271] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6da5ff64-26f6-4431-952e-6445a390ce59 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.947680] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Waiting for the task: (returnval){ [ 921.947680] env[62569]: value = "task-1250388" [ 921.947680] env[62569]: _type = "Task" [ 921.947680] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.955825] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Task: {'id': task-1250388, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.991273] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.009347] env[62569]: ERROR nova.scheduler.client.report [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [req-7a4b75c7-9eda-4bb3-a3ec-ef861d78ae2f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID fa06556a-5785-4014-b8bd-bc240a0cf716. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-7a4b75c7-9eda-4bb3-a3ec-ef861d78ae2f"}]} [ 922.025392] env[62569]: DEBUG nova.scheduler.client.report [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Refreshing inventories for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 922.039498] env[62569]: DEBUG nova.scheduler.client.report [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Updating ProviderTree inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 922.039699] env[62569]: DEBUG nova.compute.provider_tree [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 922.051110] env[62569]: DEBUG nova.scheduler.client.report [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Refreshing aggregate associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, aggregates: None {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 922.072279] env[62569]: DEBUG nova.scheduler.client.report [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Refreshing trait associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 922.120113] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 922.122406] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 922.122464] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Starting heal instance info cache {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 922.126565] env[62569]: INFO nova.compute.manager [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Took 19.86 seconds to build instance. [ 922.132712] env[62569]: DEBUG nova.compute.manager [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 922.164159] env[62569]: DEBUG nova.virt.hardware [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 922.164539] env[62569]: DEBUG nova.virt.hardware [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 922.164623] env[62569]: DEBUG nova.virt.hardware [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 922.164906] env[62569]: DEBUG nova.virt.hardware [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 922.164964] env[62569]: DEBUG nova.virt.hardware [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 922.165110] env[62569]: DEBUG nova.virt.hardware [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 922.165338] env[62569]: DEBUG nova.virt.hardware [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 922.165504] env[62569]: DEBUG nova.virt.hardware [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 922.165675] env[62569]: DEBUG nova.virt.hardware [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 922.165842] env[62569]: DEBUG nova.virt.hardware [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 922.166030] env[62569]: DEBUG nova.virt.hardware [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 922.166894] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-768d4661-0850-4c2b-ab75-bf727443d915 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.179097] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d81c36b6-02cc-4c19-83ec-3e5355b89ab2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.187848] env[62569]: DEBUG oslo_vmware.api [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250386, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.291742] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fbd1114-8406-46d6-8c70-6622cd1bd3e1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.300511] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fa3f08e-f686-40c4-b431-7ce6f4981e0a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.329889] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d1020a2-e3d2-4922-ac88-6dd7cd447c9b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.338997] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7744bd8-57e7-4657-9c5e-ee8f25263962 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.345483] env[62569]: DEBUG nova.compute.manager [req-1fef5df5-d6df-4522-8dee-e6ad1f1f293e req-9d77b6fc-72e3-48c6-9b58-aca9f7f351e4 service nova] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Received event network-vif-deleted-3d532085-b731-4296-adce-01e80852415b {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 922.355848] env[62569]: DEBUG nova.compute.provider_tree [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 922.458037] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Task: {'id': task-1250388, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.628864] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c7c60a2a-a8cf-4aa4-85f2-9dc4f215e248 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "6ccb78f3-d7f9-4090-9641-4633904efab8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.381s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.649648] env[62569]: DEBUG nova.network.neutron [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Successfully updated port: f9665162-0615-45cd-b3c2-32be38c9af60 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 922.682832] env[62569]: DEBUG oslo_vmware.api [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250386, 'name': RemoveSnapshot_Task, 'duration_secs': 1.026647} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.683142] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Deleted Snapshot of the VM instance {{(pid=62569) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 922.683393] env[62569]: INFO nova.compute.manager [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Took 15.02 seconds to snapshot the instance on the hypervisor. [ 922.887711] env[62569]: DEBUG nova.scheduler.client.report [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Updated inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with generation 116 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 922.887996] env[62569]: DEBUG nova.compute.provider_tree [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Updating resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 generation from 116 to 117 during operation: update_inventory {{(pid=62569) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 922.888239] env[62569]: DEBUG nova.compute.provider_tree [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 922.959219] env[62569]: DEBUG oslo_vmware.api [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Task: {'id': task-1250388, 'name': PowerOnVM_Task, 'duration_secs': 0.521608} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.959544] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 922.959795] env[62569]: INFO nova.compute.manager [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Took 7.01 seconds to spawn the instance on the hypervisor. [ 922.960025] env[62569]: DEBUG nova.compute.manager [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 922.961085] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6395b8aa-5201-44cb-9301-b361da849b83 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.050079] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "6ccb78f3-d7f9-4090-9641-4633904efab8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.050382] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "6ccb78f3-d7f9-4090-9641-4633904efab8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.050608] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "6ccb78f3-d7f9-4090-9641-4633904efab8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.050801] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "6ccb78f3-d7f9-4090-9641-4633904efab8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.050973] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "6ccb78f3-d7f9-4090-9641-4633904efab8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.053160] env[62569]: INFO nova.compute.manager [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Terminating instance [ 923.153091] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "refresh_cache-98089a34-074b-4bdb-92ae-f9e23d2551f1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.153291] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquired lock "refresh_cache-98089a34-074b-4bdb-92ae-f9e23d2551f1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.153472] env[62569]: DEBUG nova.network.neutron [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 923.228451] env[62569]: DEBUG nova.compute.manager [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Found 3 images (rotation: 2) {{(pid=62569) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4923}} [ 923.228668] env[62569]: DEBUG nova.compute.manager [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Rotating out 1 backups {{(pid=62569) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4931}} [ 923.228836] env[62569]: DEBUG nova.compute.manager [None req-18f395b9-3d48-4161-ac56-f6fda23a82a8 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Deleting image 870dec0b-8c00-43e0-870e-316f6bd64e1e {{(pid=62569) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4936}} [ 923.393590] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.287s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.394147] env[62569]: DEBUG nova.compute.manager [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 923.396822] env[62569]: DEBUG oslo_concurrency.lockutils [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.835s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.397017] env[62569]: DEBUG oslo_concurrency.lockutils [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.400862] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.283s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.402483] env[62569]: INFO nova.compute.claims [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 923.426164] env[62569]: INFO nova.scheduler.client.report [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Deleted allocations for instance 31ce29fa-4f60-4404-b830-21ad196f78b5 [ 923.477350] env[62569]: INFO nova.compute.manager [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Took 21.13 seconds to build instance. [ 923.558034] env[62569]: DEBUG nova.compute.manager [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 923.558034] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 923.558743] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8abad2e7-5bb9-4c89-8c24-87fa3d4d9c06 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.566429] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 923.566676] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec7115ec-d318-40c3-b24d-ac8f9ff86bd2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.572587] env[62569]: DEBUG oslo_vmware.api [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 923.572587] env[62569]: value = "task-1250389" [ 923.572587] env[62569]: _type = "Task" [ 923.572587] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.581033] env[62569]: DEBUG oslo_vmware.api [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250389, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.685756] env[62569]: DEBUG nova.network.neutron [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 923.824103] env[62569]: DEBUG nova.network.neutron [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Updating instance_info_cache with network_info: [{"id": "f9665162-0615-45cd-b3c2-32be38c9af60", "address": "fa:16:3e:2b:30:fe", "network": {"id": "ba6b71b1-bd7b-493e-8682-68d842fa935f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1892479706-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e9ea55b00b574a7db0530efcb498ca2b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9665162-06", "ovs_interfaceid": "f9665162-0615-45cd-b3c2-32be38c9af60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.907988] env[62569]: DEBUG nova.compute.utils [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 923.910134] env[62569]: DEBUG nova.compute.manager [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 923.910295] env[62569]: DEBUG nova.network.neutron [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 923.935029] env[62569]: DEBUG oslo_concurrency.lockutils [None req-00ceba58-d416-404b-883d-4c6b7db2c706 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "31ce29fa-4f60-4404-b830-21ad196f78b5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.835s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.951485] env[62569]: DEBUG nova.policy [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1418d7ccc2a4324aeb586f89c2f817b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4ba7e96d226942bd99ab29fe703c6fea', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 923.979828] env[62569]: DEBUG oslo_concurrency.lockutils [None req-75bf2a03-3e68-43e1-8b79-35f8d4bf2ea3 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Lock "f751dfdb-e950-4821-b70d-4db78e03b9b7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.651s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.088119] env[62569]: DEBUG oslo_vmware.api [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250389, 'name': PowerOffVM_Task, 'duration_secs': 0.15947} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.088119] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 924.088119] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 924.088119] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d31d9a1a-3cd3-4e54-a619-f84ab4aa37c4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.136420] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.136579] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquired lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.136728] env[62569]: DEBUG nova.network.neutron [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Forcefully refreshing network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 924.148769] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 924.148937] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 924.149205] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Deleting the datastore file [datastore2] 6ccb78f3-d7f9-4090-9641-4633904efab8 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 924.150185] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5d945b91-d34a-4e49-8d8d-b303a4197baa {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.156309] env[62569]: DEBUG oslo_vmware.api [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 924.156309] env[62569]: value = "task-1250391" [ 924.156309] env[62569]: _type = "Task" [ 924.156309] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.166069] env[62569]: DEBUG oslo_vmware.api [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250391, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.229524] env[62569]: DEBUG nova.network.neutron [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Successfully created port: c967324f-ab14-480b-a8b8-710db5827863 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 924.327634] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Releasing lock "refresh_cache-98089a34-074b-4bdb-92ae-f9e23d2551f1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.327972] env[62569]: DEBUG nova.compute.manager [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Instance network_info: |[{"id": "f9665162-0615-45cd-b3c2-32be38c9af60", "address": "fa:16:3e:2b:30:fe", "network": {"id": "ba6b71b1-bd7b-493e-8682-68d842fa935f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1892479706-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e9ea55b00b574a7db0530efcb498ca2b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9665162-06", "ovs_interfaceid": "f9665162-0615-45cd-b3c2-32be38c9af60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 924.328728] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:30:fe', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8abee039-d93e-48a7-8911-6416a3e1ff30', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f9665162-0615-45cd-b3c2-32be38c9af60', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 924.336447] env[62569]: DEBUG oslo.service.loopingcall [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 924.336984] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 924.337665] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bf79efc8-7f01-4cff-8221-21b947854cae {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.359662] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 924.359662] env[62569]: value = "task-1250392" [ 924.359662] env[62569]: _type = "Task" [ 924.359662] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.367726] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250392, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.370466] env[62569]: DEBUG nova.compute.manager [req-1612ba6e-1021-4ace-87c8-987b8ae56e69 req-82abeda9-ff02-417b-a272-5ff6d488887b service nova] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Received event network-vif-plugged-f9665162-0615-45cd-b3c2-32be38c9af60 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 924.370466] env[62569]: DEBUG oslo_concurrency.lockutils [req-1612ba6e-1021-4ace-87c8-987b8ae56e69 req-82abeda9-ff02-417b-a272-5ff6d488887b service nova] Acquiring lock "98089a34-074b-4bdb-92ae-f9e23d2551f1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.370466] env[62569]: DEBUG oslo_concurrency.lockutils [req-1612ba6e-1021-4ace-87c8-987b8ae56e69 req-82abeda9-ff02-417b-a272-5ff6d488887b service nova] Lock "98089a34-074b-4bdb-92ae-f9e23d2551f1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.370466] env[62569]: DEBUG oslo_concurrency.lockutils [req-1612ba6e-1021-4ace-87c8-987b8ae56e69 req-82abeda9-ff02-417b-a272-5ff6d488887b service nova] Lock "98089a34-074b-4bdb-92ae-f9e23d2551f1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.370671] env[62569]: DEBUG nova.compute.manager [req-1612ba6e-1021-4ace-87c8-987b8ae56e69 req-82abeda9-ff02-417b-a272-5ff6d488887b service nova] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] No waiting events found dispatching network-vif-plugged-f9665162-0615-45cd-b3c2-32be38c9af60 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 924.370709] env[62569]: WARNING nova.compute.manager [req-1612ba6e-1021-4ace-87c8-987b8ae56e69 req-82abeda9-ff02-417b-a272-5ff6d488887b service nova] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Received unexpected event network-vif-plugged-f9665162-0615-45cd-b3c2-32be38c9af60 for instance with vm_state building and task_state spawning. [ 924.370906] env[62569]: DEBUG nova.compute.manager [req-1612ba6e-1021-4ace-87c8-987b8ae56e69 req-82abeda9-ff02-417b-a272-5ff6d488887b service nova] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Received event network-changed-f9665162-0615-45cd-b3c2-32be38c9af60 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 924.371026] env[62569]: DEBUG nova.compute.manager [req-1612ba6e-1021-4ace-87c8-987b8ae56e69 req-82abeda9-ff02-417b-a272-5ff6d488887b service nova] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Refreshing instance network info cache due to event network-changed-f9665162-0615-45cd-b3c2-32be38c9af60. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 924.371228] env[62569]: DEBUG oslo_concurrency.lockutils [req-1612ba6e-1021-4ace-87c8-987b8ae56e69 req-82abeda9-ff02-417b-a272-5ff6d488887b service nova] Acquiring lock "refresh_cache-98089a34-074b-4bdb-92ae-f9e23d2551f1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.371375] env[62569]: DEBUG oslo_concurrency.lockutils [req-1612ba6e-1021-4ace-87c8-987b8ae56e69 req-82abeda9-ff02-417b-a272-5ff6d488887b service nova] Acquired lock "refresh_cache-98089a34-074b-4bdb-92ae-f9e23d2551f1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.371533] env[62569]: DEBUG nova.network.neutron [req-1612ba6e-1021-4ace-87c8-987b8ae56e69 req-82abeda9-ff02-417b-a272-5ff6d488887b service nova] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Refreshing network info cache for port f9665162-0615-45cd-b3c2-32be38c9af60 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 924.411312] env[62569]: DEBUG nova.compute.manager [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 924.582239] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquiring lock "337d88a2-30b9-4846-929e-042bd7a64a65" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.582518] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "337d88a2-30b9-4846-929e-042bd7a64a65" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.582751] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquiring lock "337d88a2-30b9-4846-929e-042bd7a64a65-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.582950] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "337d88a2-30b9-4846-929e-042bd7a64a65-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.583139] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "337d88a2-30b9-4846-929e-042bd7a64a65-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.585222] env[62569]: INFO nova.compute.manager [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Terminating instance [ 924.641994] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c699a2b4-998d-4b8c-be5a-c191d9fcb8f9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.649706] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4901a173-d744-4085-b469-3dabe833b553 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.685356] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ded36f9d-4adc-471b-8eb1-27c2fde5fd6f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.696577] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e641801-8429-4c31-baa3-4440b1d68ee2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.700744] env[62569]: DEBUG oslo_vmware.api [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250391, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138763} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.701014] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 924.701240] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 924.701444] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 924.701621] env[62569]: INFO nova.compute.manager [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Took 1.14 seconds to destroy the instance on the hypervisor. [ 924.702327] env[62569]: DEBUG oslo.service.loopingcall [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 924.702421] env[62569]: DEBUG nova.compute.manager [-] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 924.702510] env[62569]: DEBUG nova.network.neutron [-] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 924.713714] env[62569]: DEBUG nova.compute.provider_tree [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 924.786446] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Acquiring lock "f751dfdb-e950-4821-b70d-4db78e03b9b7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.786760] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Lock "f751dfdb-e950-4821-b70d-4db78e03b9b7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.787111] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Acquiring lock "f751dfdb-e950-4821-b70d-4db78e03b9b7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.787190] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Lock "f751dfdb-e950-4821-b70d-4db78e03b9b7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.787354] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Lock "f751dfdb-e950-4821-b70d-4db78e03b9b7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.789627] env[62569]: INFO nova.compute.manager [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Terminating instance [ 924.869647] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250392, 'name': CreateVM_Task} progress is 25%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.094770] env[62569]: DEBUG nova.compute.manager [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 925.094770] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 925.095709] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c057ee3f-bf3d-450a-94d7-85d2ff27b02d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.103694] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 925.103961] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cec7d8ed-26fb-4add-8870-87a6852bc1a6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.110192] env[62569]: DEBUG oslo_vmware.api [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 925.110192] env[62569]: value = "task-1250393" [ 925.110192] env[62569]: _type = "Task" [ 925.110192] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.121015] env[62569]: DEBUG oslo_vmware.api [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250393, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.261948] env[62569]: DEBUG nova.scheduler.client.report [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Updated inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with generation 117 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 925.262162] env[62569]: DEBUG nova.compute.provider_tree [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Updating resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 generation from 117 to 118 during operation: update_inventory {{(pid=62569) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 925.262359] env[62569]: DEBUG nova.compute.provider_tree [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 925.284105] env[62569]: DEBUG nova.network.neutron [req-1612ba6e-1021-4ace-87c8-987b8ae56e69 req-82abeda9-ff02-417b-a272-5ff6d488887b service nova] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Updated VIF entry in instance network info cache for port f9665162-0615-45cd-b3c2-32be38c9af60. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 925.284498] env[62569]: DEBUG nova.network.neutron [req-1612ba6e-1021-4ace-87c8-987b8ae56e69 req-82abeda9-ff02-417b-a272-5ff6d488887b service nova] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Updating instance_info_cache with network_info: [{"id": "f9665162-0615-45cd-b3c2-32be38c9af60", "address": "fa:16:3e:2b:30:fe", "network": {"id": "ba6b71b1-bd7b-493e-8682-68d842fa935f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1892479706-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e9ea55b00b574a7db0530efcb498ca2b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9665162-06", "ovs_interfaceid": "f9665162-0615-45cd-b3c2-32be38c9af60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.292988] env[62569]: DEBUG nova.compute.manager [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 925.293294] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 925.294363] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2af774f8-9286-472c-95d3-8e6f66d832cb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.304641] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 925.307437] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2ceaed8b-dfe7-40aa-8499-20f81916f192 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.314282] env[62569]: DEBUG oslo_vmware.api [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Waiting for the task: (returnval){ [ 925.314282] env[62569]: value = "task-1250394" [ 925.314282] env[62569]: _type = "Task" [ 925.314282] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.324420] env[62569]: DEBUG oslo_vmware.api [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Task: {'id': task-1250394, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.371893] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250392, 'name': CreateVM_Task, 'duration_secs': 0.741301} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.372104] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 925.373768] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.373768] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.373930] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 925.374116] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14f24e79-177b-4e49-8049-e54b8b4542c1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.379537] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 925.379537] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5210a87e-dc7f-2f0e-3f87-68037a5dd73b" [ 925.379537] env[62569]: _type = "Task" [ 925.379537] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.387592] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5210a87e-dc7f-2f0e-3f87-68037a5dd73b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.388611] env[62569]: DEBUG nova.network.neutron [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Updating instance_info_cache with network_info: [{"id": "cfddbae7-9ab0-4bc6-aea1-46a0cde743fc", "address": "fa:16:3e:b7:d6:80", "network": {"id": "406dd658-8168-4e73-9b97-861929bfaa2e", "bridge": null, "label": "tempest-ServersNegativeTestJSON-693938985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de8605118d744a93bce54e897aa849b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapcfddbae7-9a", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.422020] env[62569]: DEBUG nova.compute.manager [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 925.449163] env[62569]: DEBUG nova.virt.hardware [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 925.449443] env[62569]: DEBUG nova.virt.hardware [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 925.449686] env[62569]: DEBUG nova.virt.hardware [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 925.449979] env[62569]: DEBUG nova.virt.hardware [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 925.450213] env[62569]: DEBUG nova.virt.hardware [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 925.450392] env[62569]: DEBUG nova.virt.hardware [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 925.450693] env[62569]: DEBUG nova.virt.hardware [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 925.450880] env[62569]: DEBUG nova.virt.hardware [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 925.451073] env[62569]: DEBUG nova.virt.hardware [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 925.451299] env[62569]: DEBUG nova.virt.hardware [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 925.451491] env[62569]: DEBUG nova.virt.hardware [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 925.452411] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8cc9b86-1f7f-4a91-b114-97f9c89b5a0c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.460837] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f9ef718-cb5d-4e2a-a31b-96fa9cd25e10 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.476658] env[62569]: DEBUG nova.network.neutron [-] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.624723] env[62569]: DEBUG oslo_vmware.api [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250393, 'name': PowerOffVM_Task, 'duration_secs': 0.34178} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.624935] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 925.625124] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 925.625397] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ac9f5699-2294-4e11-ad49-ccaf929844a7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.717383] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 925.717616] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 925.717814] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Deleting the datastore file [datastore2] 337d88a2-30b9-4846-929e-042bd7a64a65 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 925.718099] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-548a9a1b-aac4-4542-871b-01bc7ddd4b49 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.724801] env[62569]: DEBUG oslo_vmware.api [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for the task: (returnval){ [ 925.724801] env[62569]: value = "task-1250396" [ 925.724801] env[62569]: _type = "Task" [ 925.724801] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.732843] env[62569]: DEBUG oslo_vmware.api [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250396, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.752274] env[62569]: DEBUG nova.network.neutron [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Successfully updated port: c967324f-ab14-480b-a8b8-710db5827863 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 925.767684] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.367s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.768356] env[62569]: DEBUG nova.compute.manager [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 925.771107] env[62569]: DEBUG oslo_concurrency.lockutils [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.213s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.771627] env[62569]: DEBUG nova.objects.instance [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lazy-loading 'resources' on Instance uuid b6e5eefc-8c06-445b-a3af-9404578b6179 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.787207] env[62569]: DEBUG oslo_concurrency.lockutils [req-1612ba6e-1021-4ace-87c8-987b8ae56e69 req-82abeda9-ff02-417b-a272-5ff6d488887b service nova] Releasing lock "refresh_cache-98089a34-074b-4bdb-92ae-f9e23d2551f1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.823801] env[62569]: DEBUG oslo_vmware.api [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Task: {'id': task-1250394, 'name': PowerOffVM_Task, 'duration_secs': 0.189037} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.824108] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 925.824308] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 925.824596] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b6ca6148-a057-4049-aa46-47e917ec93d4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.887078] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 925.887321] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 925.887513] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Deleting the datastore file [datastore2] f751dfdb-e950-4821-b70d-4db78e03b9b7 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 925.887773] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-94c5b8b7-7dff-4d6e-8dd0-77dd2268424b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.893310] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Releasing lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.893489] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Updated the network info_cache for instance {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10379}} [ 925.893728] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5210a87e-dc7f-2f0e-3f87-68037a5dd73b, 'name': SearchDatastore_Task, 'duration_secs': 0.009653} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.894292] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 925.894677] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.894904] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 925.895157] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.895311] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.895492] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 925.895757] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 925.895992] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-40e79cd9-cecb-41bd-8c85-edb42229b3be {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.898640] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 925.898960] env[62569]: DEBUG oslo_vmware.api [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Waiting for the task: (returnval){ [ 925.898960] env[62569]: value = "task-1250398" [ 925.898960] env[62569]: _type = "Task" [ 925.898960] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.899154] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 925.899761] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 925.902669] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 925.902828] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62569) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 925.903024] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 925.905466] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 925.905660] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 925.909245] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ebb52d19-5801-44cd-98ca-aa19235689c2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.911425] env[62569]: DEBUG oslo_vmware.api [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Task: {'id': task-1250398, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.915148] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 925.915148] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]522ddaf2-cec3-6e34-e5bd-9b0668ca3904" [ 925.915148] env[62569]: _type = "Task" [ 925.915148] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.924683] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]522ddaf2-cec3-6e34-e5bd-9b0668ca3904, 'name': SearchDatastore_Task, 'duration_secs': 0.007498} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.925543] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6acbdac-af33-4a1a-9074-17c2fcacc9f9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.930322] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 925.930322] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5237eab7-15e3-ee64-d44d-170fb17db487" [ 925.930322] env[62569]: _type = "Task" [ 925.930322] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.939157] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5237eab7-15e3-ee64-d44d-170fb17db487, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.979551] env[62569]: INFO nova.compute.manager [-] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Took 1.28 seconds to deallocate network for instance. [ 926.052765] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5c2608b3-808a-4576-8ae4-9eef95086706 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "b47cd2d7-0cd2-41af-8ed1-a6dfca323516" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.053071] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5c2608b3-808a-4576-8ae4-9eef95086706 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "b47cd2d7-0cd2-41af-8ed1-a6dfca323516" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.053282] env[62569]: DEBUG nova.compute.manager [None req-5c2608b3-808a-4576-8ae4-9eef95086706 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 926.054174] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ce8b303-07ac-4950-8ccd-d05251911dd0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.060934] env[62569]: DEBUG nova.compute.manager [None req-5c2608b3-808a-4576-8ae4-9eef95086706 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62569) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 926.061531] env[62569]: DEBUG nova.objects.instance [None req-5c2608b3-808a-4576-8ae4-9eef95086706 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lazy-loading 'flavor' on Instance uuid b47cd2d7-0cd2-41af-8ed1-a6dfca323516 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.234510] env[62569]: DEBUG oslo_vmware.api [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Task: {'id': task-1250396, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126026} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.234947] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 926.234947] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 926.235088] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 926.235306] env[62569]: INFO nova.compute.manager [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Took 1.14 seconds to destroy the instance on the hypervisor. [ 926.235552] env[62569]: DEBUG oslo.service.loopingcall [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.235752] env[62569]: DEBUG nova.compute.manager [-] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 926.235851] env[62569]: DEBUG nova.network.neutron [-] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 926.257962] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "refresh_cache-09eeb937-aabc-418d-9fd3-3c7a72f4ef75" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.258110] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired lock "refresh_cache-09eeb937-aabc-418d-9fd3-3c7a72f4ef75" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.258286] env[62569]: DEBUG nova.network.neutron [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 926.274118] env[62569]: DEBUG nova.objects.instance [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lazy-loading 'numa_topology' on Instance uuid b6e5eefc-8c06-445b-a3af-9404578b6179 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.279028] env[62569]: DEBUG nova.compute.utils [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 926.279028] env[62569]: DEBUG nova.compute.manager [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 926.279028] env[62569]: DEBUG nova.network.neutron [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 926.357304] env[62569]: DEBUG nova.policy [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5da4f6b7a6784a73bd3fed04b275041b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bdba6022e3d4697a336ca28ca4eccec', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 926.408272] env[62569]: DEBUG nova.compute.manager [req-7525d5b6-be4e-4ae1-b66e-e34cecff64b3 req-cb4a975e-2144-4119-8384-4dbd3e17af3e service nova] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Received event network-vif-deleted-37c1cdc3-9e78-4f21-b75e-219d6166d47b {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 926.408495] env[62569]: DEBUG nova.compute.manager [req-7525d5b6-be4e-4ae1-b66e-e34cecff64b3 req-cb4a975e-2144-4119-8384-4dbd3e17af3e service nova] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Received event network-vif-plugged-c967324f-ab14-480b-a8b8-710db5827863 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 926.408686] env[62569]: DEBUG oslo_concurrency.lockutils [req-7525d5b6-be4e-4ae1-b66e-e34cecff64b3 req-cb4a975e-2144-4119-8384-4dbd3e17af3e service nova] Acquiring lock "09eeb937-aabc-418d-9fd3-3c7a72f4ef75-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.408891] env[62569]: DEBUG oslo_concurrency.lockutils [req-7525d5b6-be4e-4ae1-b66e-e34cecff64b3 req-cb4a975e-2144-4119-8384-4dbd3e17af3e service nova] Lock "09eeb937-aabc-418d-9fd3-3c7a72f4ef75-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.409302] env[62569]: DEBUG oslo_concurrency.lockutils [req-7525d5b6-be4e-4ae1-b66e-e34cecff64b3 req-cb4a975e-2144-4119-8384-4dbd3e17af3e service nova] Lock "09eeb937-aabc-418d-9fd3-3c7a72f4ef75-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.409876] env[62569]: DEBUG nova.compute.manager [req-7525d5b6-be4e-4ae1-b66e-e34cecff64b3 req-cb4a975e-2144-4119-8384-4dbd3e17af3e service nova] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] No waiting events found dispatching network-vif-plugged-c967324f-ab14-480b-a8b8-710db5827863 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 926.409876] env[62569]: WARNING nova.compute.manager [req-7525d5b6-be4e-4ae1-b66e-e34cecff64b3 req-cb4a975e-2144-4119-8384-4dbd3e17af3e service nova] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Received unexpected event network-vif-plugged-c967324f-ab14-480b-a8b8-710db5827863 for instance with vm_state building and task_state spawning. [ 926.410388] env[62569]: DEBUG nova.compute.manager [req-7525d5b6-be4e-4ae1-b66e-e34cecff64b3 req-cb4a975e-2144-4119-8384-4dbd3e17af3e service nova] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Received event network-changed-c967324f-ab14-480b-a8b8-710db5827863 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 926.410744] env[62569]: DEBUG nova.compute.manager [req-7525d5b6-be4e-4ae1-b66e-e34cecff64b3 req-cb4a975e-2144-4119-8384-4dbd3e17af3e service nova] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Refreshing instance network info cache due to event network-changed-c967324f-ab14-480b-a8b8-710db5827863. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 926.410841] env[62569]: DEBUG oslo_concurrency.lockutils [req-7525d5b6-be4e-4ae1-b66e-e34cecff64b3 req-cb4a975e-2144-4119-8384-4dbd3e17af3e service nova] Acquiring lock "refresh_cache-09eeb937-aabc-418d-9fd3-3c7a72f4ef75" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.415891] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.416194] env[62569]: DEBUG oslo_vmware.api [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Task: {'id': task-1250398, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.124091} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.416427] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 926.416608] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 926.416789] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 926.417147] env[62569]: INFO nova.compute.manager [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Took 1.12 seconds to destroy the instance on the hypervisor. [ 926.417428] env[62569]: DEBUG oslo.service.loopingcall [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.417902] env[62569]: DEBUG nova.compute.manager [-] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 926.418014] env[62569]: DEBUG nova.network.neutron [-] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 926.444738] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5237eab7-15e3-ee64-d44d-170fb17db487, 'name': SearchDatastore_Task, 'duration_secs': 0.007787} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.445021] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.445296] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 98089a34-074b-4bdb-92ae-f9e23d2551f1/98089a34-074b-4bdb-92ae-f9e23d2551f1.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 926.445568] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c77908ed-d083-4b4b-bf24-2bebe180f0d7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.452351] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 926.452351] env[62569]: value = "task-1250399" [ 926.452351] env[62569]: _type = "Task" [ 926.452351] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.460402] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250399, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.489269] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.778390] env[62569]: DEBUG nova.objects.base [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62569) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 926.785025] env[62569]: DEBUG nova.compute.manager [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 926.829995] env[62569]: DEBUG nova.network.neutron [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Successfully created port: 9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 926.833238] env[62569]: DEBUG nova.network.neutron [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 926.838927] env[62569]: DEBUG nova.compute.manager [req-95cbf5e0-31d2-49b8-92c2-1e4258c0ef10 req-883a083d-c313-49f8-82f2-dadedce2c4e8 service nova] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Received event network-vif-deleted-89885c41-7de0-49a3-8606-d500eb1308f3 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 926.839169] env[62569]: INFO nova.compute.manager [req-95cbf5e0-31d2-49b8-92c2-1e4258c0ef10 req-883a083d-c313-49f8-82f2-dadedce2c4e8 service nova] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Neutron deleted interface 89885c41-7de0-49a3-8606-d500eb1308f3; detaching it from the instance and deleting it from the info cache [ 926.839351] env[62569]: DEBUG nova.network.neutron [req-95cbf5e0-31d2-49b8-92c2-1e4258c0ef10 req-883a083d-c313-49f8-82f2-dadedce2c4e8 service nova] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.965965] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250399, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.056519] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25ceb3df-2c71-4020-83f9-a50fcfe47bb2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.068565] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04de6764-6008-4f10-b970-8ac85ba3f64f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.073021] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c2608b3-808a-4576-8ae4-9eef95086706 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 927.073021] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4b7352c0-a0fa-4b10-b983-27fa817ab84f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.107472] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8822a357-4fbd-404a-8d40-ce97bdb981ee {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.110544] env[62569]: DEBUG oslo_vmware.api [None req-5c2608b3-808a-4576-8ae4-9eef95086706 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 927.110544] env[62569]: value = "task-1250400" [ 927.110544] env[62569]: _type = "Task" [ 927.110544] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.118350] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9656dc7a-dd00-4775-8794-fa5dcb8802c5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.125592] env[62569]: DEBUG oslo_vmware.api [None req-5c2608b3-808a-4576-8ae4-9eef95086706 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250400, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.127040] env[62569]: DEBUG nova.network.neutron [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Updating instance_info_cache with network_info: [{"id": "c967324f-ab14-480b-a8b8-710db5827863", "address": "fa:16:3e:f9:c6:da", "network": {"id": "c61b7867-7897-43ae-8f5b-19834c4390ea", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-512066189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ba7e96d226942bd99ab29fe703c6fea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc967324f-ab", "ovs_interfaceid": "c967324f-ab14-480b-a8b8-710db5827863", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.136751] env[62569]: DEBUG nova.compute.provider_tree [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.264180] env[62569]: DEBUG nova.network.neutron [-] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.342730] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-82fac725-9693-4306-87e1-81bd7b0bd77c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.352793] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf989b35-36f7-4664-8b27-c2df13b6919f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.381878] env[62569]: DEBUG nova.compute.manager [req-95cbf5e0-31d2-49b8-92c2-1e4258c0ef10 req-883a083d-c313-49f8-82f2-dadedce2c4e8 service nova] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Detach interface failed, port_id=89885c41-7de0-49a3-8606-d500eb1308f3, reason: Instance f751dfdb-e950-4821-b70d-4db78e03b9b7 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 927.463174] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250399, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.5981} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.463464] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 98089a34-074b-4bdb-92ae-f9e23d2551f1/98089a34-074b-4bdb-92ae-f9e23d2551f1.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 927.463711] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 927.464018] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-705aa15d-e3de-4522-bc66-70827c514e0f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.470888] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 927.470888] env[62569]: value = "task-1250401" [ 927.470888] env[62569]: _type = "Task" [ 927.470888] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.478711] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250401, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.513152] env[62569]: DEBUG nova.network.neutron [-] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.622836] env[62569]: DEBUG oslo_vmware.api [None req-5c2608b3-808a-4576-8ae4-9eef95086706 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250400, 'name': PowerOffVM_Task, 'duration_secs': 0.179629} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.623136] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c2608b3-808a-4576-8ae4-9eef95086706 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 927.623343] env[62569]: DEBUG nova.compute.manager [None req-5c2608b3-808a-4576-8ae4-9eef95086706 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 927.624135] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5983a11f-ad42-44fa-99af-27f680ccd6de {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.639547] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Releasing lock "refresh_cache-09eeb937-aabc-418d-9fd3-3c7a72f4ef75" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.639769] env[62569]: DEBUG nova.compute.manager [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Instance network_info: |[{"id": "c967324f-ab14-480b-a8b8-710db5827863", "address": "fa:16:3e:f9:c6:da", "network": {"id": "c61b7867-7897-43ae-8f5b-19834c4390ea", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-512066189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ba7e96d226942bd99ab29fe703c6fea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc967324f-ab", "ovs_interfaceid": "c967324f-ab14-480b-a8b8-710db5827863", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 927.640646] env[62569]: DEBUG nova.scheduler.client.report [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 927.643635] env[62569]: DEBUG oslo_concurrency.lockutils [req-7525d5b6-be4e-4ae1-b66e-e34cecff64b3 req-cb4a975e-2144-4119-8384-4dbd3e17af3e service nova] Acquired lock "refresh_cache-09eeb937-aabc-418d-9fd3-3c7a72f4ef75" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.644270] env[62569]: DEBUG nova.network.neutron [req-7525d5b6-be4e-4ae1-b66e-e34cecff64b3 req-cb4a975e-2144-4119-8384-4dbd3e17af3e service nova] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Refreshing network info cache for port c967324f-ab14-480b-a8b8-710db5827863 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 927.644818] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f9:c6:da', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6685c85e-be1e-4b7b-a6cc-3e50e59b6567', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c967324f-ab14-480b-a8b8-710db5827863', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 927.652229] env[62569]: DEBUG oslo.service.loopingcall [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.652789] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 927.653022] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8666464b-7d5d-434e-858b-7baa52ba0f28 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.673659] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 927.673659] env[62569]: value = "task-1250402" [ 927.673659] env[62569]: _type = "Task" [ 927.673659] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.681720] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250402, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.767122] env[62569]: INFO nova.compute.manager [-] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Took 1.35 seconds to deallocate network for instance. [ 927.796580] env[62569]: DEBUG nova.compute.manager [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 927.830645] env[62569]: DEBUG nova.virt.hardware [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 927.830927] env[62569]: DEBUG nova.virt.hardware [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 927.831412] env[62569]: DEBUG nova.virt.hardware [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 927.831412] env[62569]: DEBUG nova.virt.hardware [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 927.831538] env[62569]: DEBUG nova.virt.hardware [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 927.831694] env[62569]: DEBUG nova.virt.hardware [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 927.831911] env[62569]: DEBUG nova.virt.hardware [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 927.832116] env[62569]: DEBUG nova.virt.hardware [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 927.832309] env[62569]: DEBUG nova.virt.hardware [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 927.832480] env[62569]: DEBUG nova.virt.hardware [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 927.832659] env[62569]: DEBUG nova.virt.hardware [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 927.833546] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06bf321c-6f88-412b-95b6-6db9cbeb379a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.841876] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-964a09d6-c92f-4f7d-86ce-1c0b59419805 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.980930] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250401, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.118865} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.981992] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 927.982374] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20e6e678-9f69-42cb-9959-c54a19927ea5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.006342] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 98089a34-074b-4bdb-92ae-f9e23d2551f1/98089a34-074b-4bdb-92ae-f9e23d2551f1.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 928.006625] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-042849b1-0b39-47de-b367-10e3dd4deedb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.021049] env[62569]: INFO nova.compute.manager [-] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Took 1.79 seconds to deallocate network for instance. [ 928.029914] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 928.029914] env[62569]: value = "task-1250403" [ 928.029914] env[62569]: _type = "Task" [ 928.029914] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.038216] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250403, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.134935] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5c2608b3-808a-4576-8ae4-9eef95086706 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "b47cd2d7-0cd2-41af-8ed1-a6dfca323516" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.082s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.156337] env[62569]: DEBUG oslo_concurrency.lockutils [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.385s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.158739] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.933s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.160194] env[62569]: INFO nova.compute.claims [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 928.184054] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250402, 'name': CreateVM_Task, 'duration_secs': 0.364573} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.184249] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 928.184981] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.185179] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.185570] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 928.185772] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dae032fe-e6ee-46a0-84bd-ab324e7d27a0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.191745] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 928.191745] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]524b60bb-d4d3-af61-fa6d-b3a75ea74b87" [ 928.191745] env[62569]: _type = "Task" [ 928.191745] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.201317] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]524b60bb-d4d3-af61-fa6d-b3a75ea74b87, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.275154] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.465595] env[62569]: DEBUG nova.compute.manager [req-bab2d3c3-718e-4cc8-bcd6-fc1569b2e187 req-a5bc048e-9755-46ec-b962-80aba1dcceec service nova] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Received event network-vif-deleted-303f46d4-6e0a-418c-b62b-17323ab645b6 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 928.466592] env[62569]: DEBUG nova.compute.manager [req-bab2d3c3-718e-4cc8-bcd6-fc1569b2e187 req-a5bc048e-9755-46ec-b962-80aba1dcceec service nova] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Received event network-vif-plugged-9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 928.466949] env[62569]: DEBUG oslo_concurrency.lockutils [req-bab2d3c3-718e-4cc8-bcd6-fc1569b2e187 req-a5bc048e-9755-46ec-b962-80aba1dcceec service nova] Acquiring lock "af26f440-3515-4be7-9a03-8a0404c152d9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.467289] env[62569]: DEBUG oslo_concurrency.lockutils [req-bab2d3c3-718e-4cc8-bcd6-fc1569b2e187 req-a5bc048e-9755-46ec-b962-80aba1dcceec service nova] Lock "af26f440-3515-4be7-9a03-8a0404c152d9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.467603] env[62569]: DEBUG oslo_concurrency.lockutils [req-bab2d3c3-718e-4cc8-bcd6-fc1569b2e187 req-a5bc048e-9755-46ec-b962-80aba1dcceec service nova] Lock "af26f440-3515-4be7-9a03-8a0404c152d9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.467890] env[62569]: DEBUG nova.compute.manager [req-bab2d3c3-718e-4cc8-bcd6-fc1569b2e187 req-a5bc048e-9755-46ec-b962-80aba1dcceec service nova] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] No waiting events found dispatching network-vif-plugged-9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 928.468465] env[62569]: WARNING nova.compute.manager [req-bab2d3c3-718e-4cc8-bcd6-fc1569b2e187 req-a5bc048e-9755-46ec-b962-80aba1dcceec service nova] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Received unexpected event network-vif-plugged-9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c for instance with vm_state building and task_state spawning. [ 928.481549] env[62569]: DEBUG nova.network.neutron [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Successfully updated port: 9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 928.532305] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.537738] env[62569]: DEBUG nova.network.neutron [req-7525d5b6-be4e-4ae1-b66e-e34cecff64b3 req-cb4a975e-2144-4119-8384-4dbd3e17af3e service nova] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Updated VIF entry in instance network info cache for port c967324f-ab14-480b-a8b8-710db5827863. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 928.538087] env[62569]: DEBUG nova.network.neutron [req-7525d5b6-be4e-4ae1-b66e-e34cecff64b3 req-cb4a975e-2144-4119-8384-4dbd3e17af3e service nova] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Updating instance_info_cache with network_info: [{"id": "c967324f-ab14-480b-a8b8-710db5827863", "address": "fa:16:3e:f9:c6:da", "network": {"id": "c61b7867-7897-43ae-8f5b-19834c4390ea", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-512066189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ba7e96d226942bd99ab29fe703c6fea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc967324f-ab", "ovs_interfaceid": "c967324f-ab14-480b-a8b8-710db5827863", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.547347] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250403, 'name': ReconfigVM_Task, 'duration_secs': 0.368143} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.547347] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 98089a34-074b-4bdb-92ae-f9e23d2551f1/98089a34-074b-4bdb-92ae-f9e23d2551f1.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 928.548024] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aee23b06-a073-49ec-b3d3-0187c1225b0b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.555011] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 928.555011] env[62569]: value = "task-1250404" [ 928.555011] env[62569]: _type = "Task" [ 928.555011] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.563332] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250404, 'name': Rename_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.669382] env[62569]: DEBUG oslo_concurrency.lockutils [None req-07582ba8-c285-41f5-8ebe-f7b66419acbe tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "b6e5eefc-8c06-445b-a3af-9404578b6179" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 32.523s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.669592] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "b6e5eefc-8c06-445b-a3af-9404578b6179" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 11.835s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.670379] env[62569]: INFO nova.compute.manager [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Unshelving [ 928.701982] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]524b60bb-d4d3-af61-fa6d-b3a75ea74b87, 'name': SearchDatastore_Task, 'duration_secs': 0.013916} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.702420] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.702663] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 928.703155] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.703155] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.703329] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 928.703608] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f77d6f7e-0fc1-470b-bb56-339014f5a3fc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.712283] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 928.712495] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 928.713214] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4026f02-8563-4d90-9adf-38c5ff3856e7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.718839] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 928.718839] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5239ca42-afd2-f4d9-7fe5-83d56b241fba" [ 928.718839] env[62569]: _type = "Task" [ 928.718839] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.726533] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5239ca42-afd2-f4d9-7fe5-83d56b241fba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.984364] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "refresh_cache-af26f440-3515-4be7-9a03-8a0404c152d9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.984526] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired lock "refresh_cache-af26f440-3515-4be7-9a03-8a0404c152d9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.984723] env[62569]: DEBUG nova.network.neutron [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 929.042320] env[62569]: DEBUG oslo_concurrency.lockutils [req-7525d5b6-be4e-4ae1-b66e-e34cecff64b3 req-cb4a975e-2144-4119-8384-4dbd3e17af3e service nova] Releasing lock "refresh_cache-09eeb937-aabc-418d-9fd3-3c7a72f4ef75" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.065381] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250404, 'name': Rename_Task, 'duration_secs': 0.136934} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.065674] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 929.065929] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-95cbaebb-5286-4b95-aca0-9ca3eef0558d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.072053] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 929.072053] env[62569]: value = "task-1250405" [ 929.072053] env[62569]: _type = "Task" [ 929.072053] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.079580] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250405, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.233041] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5239ca42-afd2-f4d9-7fe5-83d56b241fba, 'name': SearchDatastore_Task, 'duration_secs': 0.009538} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.233853] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37c0c01d-bbd0-4bd7-b16d-c277ab20384d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.246466] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 929.246466] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52364fa9-0a22-d9a1-38dc-8544e64b94af" [ 929.246466] env[62569]: _type = "Task" [ 929.246466] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.258256] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52364fa9-0a22-d9a1-38dc-8544e64b94af, 'name': SearchDatastore_Task, 'duration_secs': 0.009689} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.258527] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.258791] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 09eeb937-aabc-418d-9fd3-3c7a72f4ef75/09eeb937-aabc-418d-9fd3-3c7a72f4ef75.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 929.259078] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0b1c9387-7a29-4ec4-9741-4cab523e5306 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.266554] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 929.266554] env[62569]: value = "task-1250406" [ 929.266554] env[62569]: _type = "Task" [ 929.266554] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.274820] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250406, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.408129] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b66ed6-2ae6-4ae6-b1b1-743d225a1553 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.418380] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4478f2a7-da53-4820-9f08-f3f92642ab8f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.453397] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d304bdc-4481-4efe-a85d-d2faa3e709d0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.464191] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea14c421-b7a7-4297-8c5a-aeb16a55d015 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.481380] env[62569]: DEBUG nova.compute.provider_tree [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.535077] env[62569]: DEBUG nova.network.neutron [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 929.586528] env[62569]: DEBUG oslo_vmware.api [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250405, 'name': PowerOnVM_Task, 'duration_secs': 0.485976} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.586835] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 929.587099] env[62569]: INFO nova.compute.manager [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Took 7.45 seconds to spawn the instance on the hypervisor. [ 929.587304] env[62569]: DEBUG nova.compute.manager [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 929.588084] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-546d122c-7938-4611-901e-65be82ff3174 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.649549] env[62569]: DEBUG nova.compute.manager [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Stashing vm_state: stopped {{(pid=62569) _prep_resize /opt/stack/nova/nova/compute/manager.py:5997}} [ 929.698699] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.776285] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250406, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.470949} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.776775] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 09eeb937-aabc-418d-9fd3-3c7a72f4ef75/09eeb937-aabc-418d-9fd3-3c7a72f4ef75.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 929.776775] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 929.777018] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a15c5681-9a5d-4ed9-ad7e-4f4be6d7b78b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.783042] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 929.783042] env[62569]: value = "task-1250407" [ 929.783042] env[62569]: _type = "Task" [ 929.783042] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.791318] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250407, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.812576] env[62569]: DEBUG nova.network.neutron [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Updating instance_info_cache with network_info: [{"id": "9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c", "address": "fa:16:3e:75:9a:a6", "network": {"id": "30d3bf4a-8f55-492f-a049-181e29731818", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1244949475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bdba6022e3d4697a336ca28ca4eccec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c1cf99c-fb", "ovs_interfaceid": "9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.986044] env[62569]: DEBUG nova.scheduler.client.report [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 930.106775] env[62569]: INFO nova.compute.manager [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Took 23.70 seconds to build instance. [ 930.169906] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.292742] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250407, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063696} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.292989] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 930.293786] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbcb27cd-e2a0-4817-becb-e9e0c2b92fc1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.315828] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] 09eeb937-aabc-418d-9fd3-3c7a72f4ef75/09eeb937-aabc-418d-9fd3-3c7a72f4ef75.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 930.316391] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Releasing lock "refresh_cache-af26f440-3515-4be7-9a03-8a0404c152d9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.316714] env[62569]: DEBUG nova.compute.manager [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Instance network_info: |[{"id": "9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c", "address": "fa:16:3e:75:9a:a6", "network": {"id": "30d3bf4a-8f55-492f-a049-181e29731818", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1244949475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bdba6022e3d4697a336ca28ca4eccec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c1cf99c-fb", "ovs_interfaceid": "9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 930.316974] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a63e6d7a-9514-4013-b599-9f77780c2161 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.331455] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:75:9a:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '983826cf-6390-4ec6-bf97-30a1060947fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 930.339614] env[62569]: DEBUG oslo.service.loopingcall [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 930.339614] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 930.340295] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7b6a8874-1bbd-486d-a30a-67c45f191fb6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.355832] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 930.355832] env[62569]: value = "task-1250408" [ 930.355832] env[62569]: _type = "Task" [ 930.355832] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.361864] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 930.361864] env[62569]: value = "task-1250409" [ 930.361864] env[62569]: _type = "Task" [ 930.361864] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.366108] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250408, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.375847] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250409, 'name': CreateVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.492057] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.333s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.492900] env[62569]: DEBUG nova.compute.manager [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 930.498055] env[62569]: DEBUG nova.compute.manager [req-306cb0f0-1a55-4eb2-b143-6c4635b2da98 req-1189dff7-9627-48a3-9883-620ead9c4abe service nova] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Received event network-changed-9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 930.498055] env[62569]: DEBUG nova.compute.manager [req-306cb0f0-1a55-4eb2-b143-6c4635b2da98 req-1189dff7-9627-48a3-9883-620ead9c4abe service nova] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Refreshing instance network info cache due to event network-changed-9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 930.498055] env[62569]: DEBUG oslo_concurrency.lockutils [req-306cb0f0-1a55-4eb2-b143-6c4635b2da98 req-1189dff7-9627-48a3-9883-620ead9c4abe service nova] Acquiring lock "refresh_cache-af26f440-3515-4be7-9a03-8a0404c152d9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.498263] env[62569]: DEBUG oslo_concurrency.lockutils [req-306cb0f0-1a55-4eb2-b143-6c4635b2da98 req-1189dff7-9627-48a3-9883-620ead9c4abe service nova] Acquired lock "refresh_cache-af26f440-3515-4be7-9a03-8a0404c152d9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.498351] env[62569]: DEBUG nova.network.neutron [req-306cb0f0-1a55-4eb2-b143-6c4635b2da98 req-1189dff7-9627-48a3-9883-620ead9c4abe service nova] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Refreshing network info cache for port 9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 930.499794] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.889s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.501080] env[62569]: INFO nova.compute.claims [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 930.607919] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f09c1ac-030b-4d41-aa89-d87be4f950f7 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.209s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.866627] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250408, 'name': ReconfigVM_Task, 'duration_secs': 0.424243} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.869834] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Reconfigured VM instance instance-0000005b to attach disk [datastore2] 09eeb937-aabc-418d-9fd3-3c7a72f4ef75/09eeb937-aabc-418d-9fd3-3c7a72f4ef75.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 930.870581] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fa0a078c-a5aa-4cc9-baae-d0711cb49999 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.877223] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250409, 'name': CreateVM_Task, 'duration_secs': 0.40556} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.878634] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 930.878744] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 930.878744] env[62569]: value = "task-1250410" [ 930.878744] env[62569]: _type = "Task" [ 930.878744] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.879478] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.879923] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.880104] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 930.880463] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9140955b-8407-4578-9903-d2f353a8ca4b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.887506] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 930.887506] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5200d001-92e5-ad30-ffb8-82f35458214f" [ 930.887506] env[62569]: _type = "Task" [ 930.887506] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.893021] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250410, 'name': Rename_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.898419] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5200d001-92e5-ad30-ffb8-82f35458214f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.001451] env[62569]: DEBUG nova.compute.utils [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 931.003323] env[62569]: DEBUG nova.compute.manager [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 931.003679] env[62569]: DEBUG nova.network.neutron [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 931.153993] env[62569]: DEBUG nova.policy [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a272992a4894805bfb958680a37da8e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b20340a1ce0447cae1bfd8b7c28928f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 931.225434] env[62569]: DEBUG nova.network.neutron [req-306cb0f0-1a55-4eb2-b143-6c4635b2da98 req-1189dff7-9627-48a3-9883-620ead9c4abe service nova] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Updated VIF entry in instance network info cache for port 9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 931.225907] env[62569]: DEBUG nova.network.neutron [req-306cb0f0-1a55-4eb2-b143-6c4635b2da98 req-1189dff7-9627-48a3-9883-620ead9c4abe service nova] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Updating instance_info_cache with network_info: [{"id": "9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c", "address": "fa:16:3e:75:9a:a6", "network": {"id": "30d3bf4a-8f55-492f-a049-181e29731818", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1244949475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bdba6022e3d4697a336ca28ca4eccec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c1cf99c-fb", "ovs_interfaceid": "9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.391617] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250410, 'name': Rename_Task, 'duration_secs': 0.188307} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.395540] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 931.395882] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c2baacdb-c65c-4372-b5f8-a1fde61d6b76 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.403062] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5200d001-92e5-ad30-ffb8-82f35458214f, 'name': SearchDatastore_Task, 'duration_secs': 0.010086} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.404426] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.404903] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 931.405226] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.405448] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.405692] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 931.406045] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 931.406045] env[62569]: value = "task-1250411" [ 931.406045] env[62569]: _type = "Task" [ 931.406045] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.406336] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-17963aea-ba7c-4d64-97e8-743e8c94fd51 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.415815] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250411, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.416986] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 931.417386] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 931.418129] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0865e5a-99f9-45b1-9b07-71d362afb532 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.422827] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 931.422827] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52425773-1067-08c4-e67f-66d1ba5d6334" [ 931.422827] env[62569]: _type = "Task" [ 931.422827] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.430820] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52425773-1067-08c4-e67f-66d1ba5d6334, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.507265] env[62569]: DEBUG nova.compute.manager [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 931.732082] env[62569]: DEBUG oslo_concurrency.lockutils [req-306cb0f0-1a55-4eb2-b143-6c4635b2da98 req-1189dff7-9627-48a3-9883-620ead9c4abe service nova] Releasing lock "refresh_cache-af26f440-3515-4be7-9a03-8a0404c152d9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.762906] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c006f155-2889-46fb-874b-3cf7aaebb67a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.773277] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8ee1d3c-8713-4a1e-95a1-431bbcfd0de4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.807740] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1f9997a-6583-4dbb-accc-6a98f2be3391 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.815838] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d77a07f5-fbfa-48d4-a1ed-0e37769aa229 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.820947] env[62569]: DEBUG nova.network.neutron [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Successfully created port: e1885f70-c997-4345-8ff7-0927c947b6cf {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 931.826183] env[62569]: DEBUG nova.compute.manager [req-ef8a5e57-127d-484c-8d32-4b380130c3da req-41f73b58-6fcb-4846-806e-d79135259ebf service nova] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Received event network-changed-f9665162-0615-45cd-b3c2-32be38c9af60 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 931.826500] env[62569]: DEBUG nova.compute.manager [req-ef8a5e57-127d-484c-8d32-4b380130c3da req-41f73b58-6fcb-4846-806e-d79135259ebf service nova] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Refreshing instance network info cache due to event network-changed-f9665162-0615-45cd-b3c2-32be38c9af60. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 931.826741] env[62569]: DEBUG oslo_concurrency.lockutils [req-ef8a5e57-127d-484c-8d32-4b380130c3da req-41f73b58-6fcb-4846-806e-d79135259ebf service nova] Acquiring lock "refresh_cache-98089a34-074b-4bdb-92ae-f9e23d2551f1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.826891] env[62569]: DEBUG oslo_concurrency.lockutils [req-ef8a5e57-127d-484c-8d32-4b380130c3da req-41f73b58-6fcb-4846-806e-d79135259ebf service nova] Acquired lock "refresh_cache-98089a34-074b-4bdb-92ae-f9e23d2551f1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.827090] env[62569]: DEBUG nova.network.neutron [req-ef8a5e57-127d-484c-8d32-4b380130c3da req-41f73b58-6fcb-4846-806e-d79135259ebf service nova] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Refreshing network info cache for port f9665162-0615-45cd-b3c2-32be38c9af60 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 931.836863] env[62569]: DEBUG nova.compute.provider_tree [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 931.919056] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250411, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.932210] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52425773-1067-08c4-e67f-66d1ba5d6334, 'name': SearchDatastore_Task, 'duration_secs': 0.014819} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.933013] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa1548e4-08f4-4e22-a263-8dc5d4651b71 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.938219] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 931.938219] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ba355b-fbec-5b29-9f08-0eb342433482" [ 931.938219] env[62569]: _type = "Task" [ 931.938219] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.946029] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ba355b-fbec-5b29-9f08-0eb342433482, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.364932] env[62569]: ERROR nova.scheduler.client.report [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [req-9bdd909d-d1b4-4f32-b233-470df42e3cce] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID fa06556a-5785-4014-b8bd-bc240a0cf716. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-9bdd909d-d1b4-4f32-b233-470df42e3cce"}]} [ 932.387284] env[62569]: DEBUG nova.scheduler.client.report [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Refreshing inventories for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 932.405684] env[62569]: DEBUG nova.scheduler.client.report [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Updating ProviderTree inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 932.405937] env[62569]: DEBUG nova.compute.provider_tree [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 932.418322] env[62569]: DEBUG nova.scheduler.client.report [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Refreshing aggregate associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, aggregates: None {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 932.423425] env[62569]: DEBUG oslo_vmware.api [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250411, 'name': PowerOnVM_Task, 'duration_secs': 0.532894} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.424747] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 932.424747] env[62569]: INFO nova.compute.manager [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Took 7.00 seconds to spawn the instance on the hypervisor. [ 932.424747] env[62569]: DEBUG nova.compute.manager [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 932.424911] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e873c24-81c0-47a4-8774-ea5ed1c19553 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.447310] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ba355b-fbec-5b29-9f08-0eb342433482, 'name': SearchDatastore_Task, 'duration_secs': 0.010732} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.448478] env[62569]: DEBUG nova.scheduler.client.report [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Refreshing trait associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 932.450246] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.450917] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] af26f440-3515-4be7-9a03-8a0404c152d9/af26f440-3515-4be7-9a03-8a0404c152d9.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 932.451051] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1ea92521-2a9a-45c5-b8ab-8e76648c049b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.458768] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 932.458768] env[62569]: value = "task-1250412" [ 932.458768] env[62569]: _type = "Task" [ 932.458768] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.466418] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250412, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.517863] env[62569]: DEBUG nova.compute.manager [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 932.548719] env[62569]: DEBUG nova.virt.hardware [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 932.548987] env[62569]: DEBUG nova.virt.hardware [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 932.549178] env[62569]: DEBUG nova.virt.hardware [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 932.549373] env[62569]: DEBUG nova.virt.hardware [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 932.549525] env[62569]: DEBUG nova.virt.hardware [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 932.549704] env[62569]: DEBUG nova.virt.hardware [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 932.549947] env[62569]: DEBUG nova.virt.hardware [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 932.550135] env[62569]: DEBUG nova.virt.hardware [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 932.550316] env[62569]: DEBUG nova.virt.hardware [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 932.550487] env[62569]: DEBUG nova.virt.hardware [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 932.550670] env[62569]: DEBUG nova.virt.hardware [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 932.551653] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99eb003b-604d-45fb-9ad0-3bf846457eb8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.561986] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9cc9d04-437d-4506-a197-305b5f6f1c2c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.640675] env[62569]: DEBUG oslo_concurrency.lockutils [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.640948] env[62569]: DEBUG oslo_concurrency.lockutils [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.641228] env[62569]: DEBUG oslo_concurrency.lockutils [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.641508] env[62569]: DEBUG oslo_concurrency.lockutils [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.641716] env[62569]: DEBUG oslo_concurrency.lockutils [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.646123] env[62569]: INFO nova.compute.manager [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Terminating instance [ 932.715316] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00008846-4949-493e-8545-9eb42db2d64e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.723273] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-657b5adf-387b-4bad-9f4f-1b8f29fd7d3b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.757521] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d49362b-73ff-4f9e-b0de-7aa91c46ceaf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.766956] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47106fe9-9b4c-4428-b12e-aacc5adebb75 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.780712] env[62569]: DEBUG nova.compute.provider_tree [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 932.797749] env[62569]: DEBUG nova.network.neutron [req-ef8a5e57-127d-484c-8d32-4b380130c3da req-41f73b58-6fcb-4846-806e-d79135259ebf service nova] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Updated VIF entry in instance network info cache for port f9665162-0615-45cd-b3c2-32be38c9af60. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 932.798151] env[62569]: DEBUG nova.network.neutron [req-ef8a5e57-127d-484c-8d32-4b380130c3da req-41f73b58-6fcb-4846-806e-d79135259ebf service nova] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Updating instance_info_cache with network_info: [{"id": "f9665162-0615-45cd-b3c2-32be38c9af60", "address": "fa:16:3e:2b:30:fe", "network": {"id": "ba6b71b1-bd7b-493e-8682-68d842fa935f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1892479706-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.224", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e9ea55b00b574a7db0530efcb498ca2b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9665162-06", "ovs_interfaceid": "f9665162-0615-45cd-b3c2-32be38c9af60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.946075] env[62569]: INFO nova.compute.manager [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Took 24.41 seconds to build instance. [ 932.968713] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250412, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.149972] env[62569]: DEBUG nova.compute.manager [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 933.150250] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 933.151227] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-783cea05-0f5c-4021-9215-f9ffe9425c85 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.160499] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 933.160742] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-25e73cf3-fbdf-411e-9631-dab455b3fba8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.168200] env[62569]: DEBUG oslo_vmware.api [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 933.168200] env[62569]: value = "task-1250413" [ 933.168200] env[62569]: _type = "Task" [ 933.168200] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.180838] env[62569]: DEBUG oslo_vmware.api [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250413, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.300631] env[62569]: DEBUG oslo_concurrency.lockutils [req-ef8a5e57-127d-484c-8d32-4b380130c3da req-41f73b58-6fcb-4846-806e-d79135259ebf service nova] Releasing lock "refresh_cache-98089a34-074b-4bdb-92ae-f9e23d2551f1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.314493] env[62569]: DEBUG nova.scheduler.client.report [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Updated inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with generation 120 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 933.314863] env[62569]: DEBUG nova.compute.provider_tree [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Updating resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 generation from 120 to 121 during operation: update_inventory {{(pid=62569) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 933.315019] env[62569]: DEBUG nova.compute.provider_tree [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 933.447870] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9d75c940-38e5-44cb-bfc5-59186f2e8661 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "09eeb937-aabc-418d-9fd3-3c7a72f4ef75" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.921s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.469131] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250412, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.852965} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.469486] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] af26f440-3515-4be7-9a03-8a0404c152d9/af26f440-3515-4be7-9a03-8a0404c152d9.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 933.469715] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 933.469977] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ab95ac5e-aa73-4751-9290-d9e5ecb54bfc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.475708] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 933.475708] env[62569]: value = "task-1250414" [ 933.475708] env[62569]: _type = "Task" [ 933.475708] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.483773] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250414, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.627768] env[62569]: DEBUG nova.network.neutron [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Successfully updated port: e1885f70-c997-4345-8ff7-0927c947b6cf {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 933.678073] env[62569]: DEBUG oslo_vmware.api [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250413, 'name': PowerOffVM_Task, 'duration_secs': 0.394457} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.678496] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 933.678711] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 933.678994] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-69ba781c-4cdd-496b-b98e-49b7856c6b28 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.776803] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 933.777060] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 933.777265] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Deleting the datastore file [datastore1] bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 933.777538] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4990ecc9-f350-4c06-92df-27f414dcb357 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.783742] env[62569]: DEBUG oslo_vmware.api [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 933.783742] env[62569]: value = "task-1250416" [ 933.783742] env[62569]: _type = "Task" [ 933.783742] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.792052] env[62569]: DEBUG oslo_vmware.api [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250416, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.820166] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.320s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.821321] env[62569]: DEBUG nova.compute.manager [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 933.824274] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.833s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.824507] env[62569]: DEBUG nova.objects.instance [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lazy-loading 'resources' on Instance uuid 05db90a6-3faf-4878-b782-ed17df47ed5f {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 933.857375] env[62569]: DEBUG nova.compute.manager [req-0512f336-469a-4edf-8339-5ed9fb2653b3 req-2c63e021-7514-41c6-998a-0e6f4d6ccb7d service nova] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Received event network-vif-plugged-e1885f70-c997-4345-8ff7-0927c947b6cf {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 933.857743] env[62569]: DEBUG oslo_concurrency.lockutils [req-0512f336-469a-4edf-8339-5ed9fb2653b3 req-2c63e021-7514-41c6-998a-0e6f4d6ccb7d service nova] Acquiring lock "d1fdff70-0401-49ca-bbbb-ef5eda266c9c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.857743] env[62569]: DEBUG oslo_concurrency.lockutils [req-0512f336-469a-4edf-8339-5ed9fb2653b3 req-2c63e021-7514-41c6-998a-0e6f4d6ccb7d service nova] Lock "d1fdff70-0401-49ca-bbbb-ef5eda266c9c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.857968] env[62569]: DEBUG oslo_concurrency.lockutils [req-0512f336-469a-4edf-8339-5ed9fb2653b3 req-2c63e021-7514-41c6-998a-0e6f4d6ccb7d service nova] Lock "d1fdff70-0401-49ca-bbbb-ef5eda266c9c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.858517] env[62569]: DEBUG nova.compute.manager [req-0512f336-469a-4edf-8339-5ed9fb2653b3 req-2c63e021-7514-41c6-998a-0e6f4d6ccb7d service nova] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] No waiting events found dispatching network-vif-plugged-e1885f70-c997-4345-8ff7-0927c947b6cf {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 933.858517] env[62569]: WARNING nova.compute.manager [req-0512f336-469a-4edf-8339-5ed9fb2653b3 req-2c63e021-7514-41c6-998a-0e6f4d6ccb7d service nova] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Received unexpected event network-vif-plugged-e1885f70-c997-4345-8ff7-0927c947b6cf for instance with vm_state building and task_state spawning. [ 933.858517] env[62569]: DEBUG nova.compute.manager [req-0512f336-469a-4edf-8339-5ed9fb2653b3 req-2c63e021-7514-41c6-998a-0e6f4d6ccb7d service nova] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Received event network-changed-e1885f70-c997-4345-8ff7-0927c947b6cf {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 933.858779] env[62569]: DEBUG nova.compute.manager [req-0512f336-469a-4edf-8339-5ed9fb2653b3 req-2c63e021-7514-41c6-998a-0e6f4d6ccb7d service nova] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Refreshing instance network info cache due to event network-changed-e1885f70-c997-4345-8ff7-0927c947b6cf. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 933.858779] env[62569]: DEBUG oslo_concurrency.lockutils [req-0512f336-469a-4edf-8339-5ed9fb2653b3 req-2c63e021-7514-41c6-998a-0e6f4d6ccb7d service nova] Acquiring lock "refresh_cache-d1fdff70-0401-49ca-bbbb-ef5eda266c9c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.858932] env[62569]: DEBUG oslo_concurrency.lockutils [req-0512f336-469a-4edf-8339-5ed9fb2653b3 req-2c63e021-7514-41c6-998a-0e6f4d6ccb7d service nova] Acquired lock "refresh_cache-d1fdff70-0401-49ca-bbbb-ef5eda266c9c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.859156] env[62569]: DEBUG nova.network.neutron [req-0512f336-469a-4edf-8339-5ed9fb2653b3 req-2c63e021-7514-41c6-998a-0e6f4d6ccb7d service nova] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Refreshing network info cache for port e1885f70-c997-4345-8ff7-0927c947b6cf {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 933.986450] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250414, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066228} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.986769] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 933.987554] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6adab4ea-2dca-4500-964c-4f323c9f2168 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.009406] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] af26f440-3515-4be7-9a03-8a0404c152d9/af26f440-3515-4be7-9a03-8a0404c152d9.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 934.009662] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34505265-28ad-4a55-a153-d4d252838edd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.028687] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 934.028687] env[62569]: value = "task-1250417" [ 934.028687] env[62569]: _type = "Task" [ 934.028687] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.036121] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250417, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.130583] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "refresh_cache-d1fdff70-0401-49ca-bbbb-ef5eda266c9c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.293683] env[62569]: DEBUG oslo_vmware.api [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250416, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.331570] env[62569]: DEBUG nova.compute.utils [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 934.335669] env[62569]: DEBUG nova.compute.manager [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 934.335850] env[62569]: DEBUG nova.network.neutron [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 934.374929] env[62569]: DEBUG nova.compute.manager [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Stashing vm_state: active {{(pid=62569) _prep_resize /opt/stack/nova/nova/compute/manager.py:5997}} [ 934.402676] env[62569]: DEBUG nova.network.neutron [req-0512f336-469a-4edf-8339-5ed9fb2653b3 req-2c63e021-7514-41c6-998a-0e6f4d6ccb7d service nova] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 934.407145] env[62569]: DEBUG nova.policy [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ab6d20862c54432cbafdda33ccc974ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '530ca7e7924743ab91a362a064a3111b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 934.482078] env[62569]: DEBUG nova.network.neutron [req-0512f336-469a-4edf-8339-5ed9fb2653b3 req-2c63e021-7514-41c6-998a-0e6f4d6ccb7d service nova] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.541092] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250417, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.659320] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b86c030a-4c1c-414c-adbe-91f6bd12aa43 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.666977] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbed585b-4121-4d7c-8bee-a01f2a879768 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.700497] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a409f513-0b0c-469e-bc20-850f0b87f934 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.708495] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68413088-8a17-4915-bd64-cf1821819669 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.723937] env[62569]: DEBUG nova.compute.provider_tree [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 934.794926] env[62569]: DEBUG oslo_vmware.api [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250416, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.729131} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.795207] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 934.795408] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 934.795593] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 934.796154] env[62569]: INFO nova.compute.manager [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Took 1.65 seconds to destroy the instance on the hypervisor. [ 934.796154] env[62569]: DEBUG oslo.service.loopingcall [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 934.796331] env[62569]: DEBUG nova.compute.manager [-] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 934.796331] env[62569]: DEBUG nova.network.neutron [-] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 934.838937] env[62569]: DEBUG nova.compute.manager [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 934.894872] env[62569]: DEBUG oslo_concurrency.lockutils [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.923590] env[62569]: DEBUG nova.network.neutron [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Successfully created port: 6ca4b28a-1255-4f88-a7b6-54b0378768b1 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 934.984158] env[62569]: DEBUG oslo_concurrency.lockutils [req-0512f336-469a-4edf-8339-5ed9fb2653b3 req-2c63e021-7514-41c6-998a-0e6f4d6ccb7d service nova] Releasing lock "refresh_cache-d1fdff70-0401-49ca-bbbb-ef5eda266c9c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.984598] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquired lock "refresh_cache-d1fdff70-0401-49ca-bbbb-ef5eda266c9c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.984841] env[62569]: DEBUG nova.network.neutron [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 935.039813] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250417, 'name': ReconfigVM_Task, 'duration_secs': 0.733183} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.039813] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Reconfigured VM instance instance-0000005c to attach disk [datastore2] af26f440-3515-4be7-9a03-8a0404c152d9/af26f440-3515-4be7-9a03-8a0404c152d9.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 935.040378] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ce478ea3-9c85-4807-9623-cfb5e56b9aab {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.046873] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 935.046873] env[62569]: value = "task-1250418" [ 935.046873] env[62569]: _type = "Task" [ 935.046873] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.056591] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250418, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.227101] env[62569]: DEBUG nova.scheduler.client.report [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 935.308246] env[62569]: DEBUG nova.compute.manager [req-64941258-cd97-41dc-8807-510f8faa3a02 req-ce998a06-e324-44d5-8196-76bbcfb52322 service nova] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Received event network-vif-deleted-d64908ff-b5fa-42fe-8a49-1f39a27ca4b7 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 935.308246] env[62569]: INFO nova.compute.manager [req-64941258-cd97-41dc-8807-510f8faa3a02 req-ce998a06-e324-44d5-8196-76bbcfb52322 service nova] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Neutron deleted interface d64908ff-b5fa-42fe-8a49-1f39a27ca4b7; detaching it from the instance and deleting it from the info cache [ 935.308376] env[62569]: DEBUG nova.network.neutron [req-64941258-cd97-41dc-8807-510f8faa3a02 req-ce998a06-e324-44d5-8196-76bbcfb52322 service nova] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.520138] env[62569]: DEBUG nova.network.neutron [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 935.556952] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250418, 'name': Rename_Task, 'duration_secs': 0.327711} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.559473] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 935.559743] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f00dd31d-7760-440d-8301-f5c8dcfcb3eb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.565690] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 935.565690] env[62569]: value = "task-1250419" [ 935.565690] env[62569]: _type = "Task" [ 935.565690] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.573136] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250419, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.653596] env[62569]: DEBUG nova.network.neutron [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Updating instance_info_cache with network_info: [{"id": "e1885f70-c997-4345-8ff7-0927c947b6cf", "address": "fa:16:3e:bd:ca:3d", "network": {"id": "8ea5e138-813b-4c5c-88a1-b1ea79807f10", "bridge": "br-int", "label": "tempest-ImagesTestJSON-524163445-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b20340a1ce0447cae1bfd8b7c28928f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1885f70-c9", "ovs_interfaceid": "e1885f70-c997-4345-8ff7-0927c947b6cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.731999] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.908s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.735224] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 9.319s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.735440] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.735604] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62569) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 935.735918] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.247s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.736161] env[62569]: DEBUG nova.objects.instance [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lazy-loading 'resources' on Instance uuid 6ccb78f3-d7f9-4090-9641-4633904efab8 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.739167] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27dab1bd-18da-4516-8ff0-9b748ed2d890 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.748022] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6721f5bb-fc2e-4e92-beb5-7a056871d02d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.763376] env[62569]: INFO nova.scheduler.client.report [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Deleted allocations for instance 05db90a6-3faf-4878-b782-ed17df47ed5f [ 935.765345] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8c03541-5f53-4aae-90f6-418b8620b364 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.774333] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d00ad5d-ea12-4b67-a7d8-6c17e98fe821 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.804370] env[62569]: DEBUG nova.network.neutron [-] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.806368] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179413MB free_disk=146GB free_vcpus=48 pci_devices=None {{(pid=62569) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 935.806911] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.810660] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8d4311de-9fcf-46f5-9691-9b6f29c669f2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.819989] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-300b51aa-33cb-45c0-83bb-eeeb835d1d40 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.851129] env[62569]: DEBUG nova.compute.manager [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 935.853165] env[62569]: DEBUG nova.compute.manager [req-64941258-cd97-41dc-8807-510f8faa3a02 req-ce998a06-e324-44d5-8196-76bbcfb52322 service nova] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Detach interface failed, port_id=d64908ff-b5fa-42fe-8a49-1f39a27ca4b7, reason: Instance bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 935.884546] env[62569]: DEBUG nova.virt.hardware [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 935.884788] env[62569]: DEBUG nova.virt.hardware [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 935.884953] env[62569]: DEBUG nova.virt.hardware [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 935.885172] env[62569]: DEBUG nova.virt.hardware [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 935.885325] env[62569]: DEBUG nova.virt.hardware [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 935.885600] env[62569]: DEBUG nova.virt.hardware [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 935.885687] env[62569]: DEBUG nova.virt.hardware [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 935.885874] env[62569]: DEBUG nova.virt.hardware [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 935.886328] env[62569]: DEBUG nova.virt.hardware [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 935.886328] env[62569]: DEBUG nova.virt.hardware [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 935.886581] env[62569]: DEBUG nova.virt.hardware [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 935.888890] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f853f359-b095-4c4b-b26a-80d4e6b4c4f3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.895614] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c95438b7-1151-4046-aa1b-06496cf1d637 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.076597] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250419, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.156931] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Releasing lock "refresh_cache-d1fdff70-0401-49ca-bbbb-ef5eda266c9c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.157222] env[62569]: DEBUG nova.compute.manager [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Instance network_info: |[{"id": "e1885f70-c997-4345-8ff7-0927c947b6cf", "address": "fa:16:3e:bd:ca:3d", "network": {"id": "8ea5e138-813b-4c5c-88a1-b1ea79807f10", "bridge": "br-int", "label": "tempest-ImagesTestJSON-524163445-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b20340a1ce0447cae1bfd8b7c28928f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f65996a3-f865-4492-9377-cd14ec8b3aae", "external-id": "nsx-vlan-transportzone-31", "segmentation_id": 31, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape1885f70-c9", "ovs_interfaceid": "e1885f70-c997-4345-8ff7-0927c947b6cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 936.157970] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bd:ca:3d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f65996a3-f865-4492-9377-cd14ec8b3aae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e1885f70-c997-4345-8ff7-0927c947b6cf', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 936.165607] env[62569]: DEBUG oslo.service.loopingcall [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 936.165840] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 936.166085] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2d59e601-82f1-4b8c-a881-b7f2935a5424 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.186040] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 936.186040] env[62569]: value = "task-1250420" [ 936.186040] env[62569]: _type = "Task" [ 936.186040] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.193983] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250420, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.274198] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3730dd0e-736d-437e-89cf-40fb61970d38 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "05db90a6-3faf-4878-b782-ed17df47ed5f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.729s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.309016] env[62569]: INFO nova.compute.manager [-] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Took 1.51 seconds to deallocate network for instance. [ 936.485489] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eb30c2e-9579-4664-9d06-34a2c116167d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.493453] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2a8030d-677c-4095-aa7a-1da1562d85be {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.535583] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49048674-6bb2-477d-a6ae-f000e685bb12 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.544207] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9188bb12-2494-4092-a81c-c1aa26077830 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.551245] env[62569]: DEBUG nova.compute.manager [req-a3635fa4-39f1-4d1d-accf-a5e5a1131d97 req-2570ef26-7e6a-4886-afb2-0b1ea9c9b6f8 service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Received event network-vif-plugged-6ca4b28a-1255-4f88-a7b6-54b0378768b1 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 936.551456] env[62569]: DEBUG oslo_concurrency.lockutils [req-a3635fa4-39f1-4d1d-accf-a5e5a1131d97 req-2570ef26-7e6a-4886-afb2-0b1ea9c9b6f8 service nova] Acquiring lock "93eb1a09-269a-44c8-8f9d-3a2b7e0d715e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.551664] env[62569]: DEBUG oslo_concurrency.lockutils [req-a3635fa4-39f1-4d1d-accf-a5e5a1131d97 req-2570ef26-7e6a-4886-afb2-0b1ea9c9b6f8 service nova] Lock "93eb1a09-269a-44c8-8f9d-3a2b7e0d715e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.551831] env[62569]: DEBUG oslo_concurrency.lockutils [req-a3635fa4-39f1-4d1d-accf-a5e5a1131d97 req-2570ef26-7e6a-4886-afb2-0b1ea9c9b6f8 service nova] Lock "93eb1a09-269a-44c8-8f9d-3a2b7e0d715e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.552008] env[62569]: DEBUG nova.compute.manager [req-a3635fa4-39f1-4d1d-accf-a5e5a1131d97 req-2570ef26-7e6a-4886-afb2-0b1ea9c9b6f8 service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] No waiting events found dispatching network-vif-plugged-6ca4b28a-1255-4f88-a7b6-54b0378768b1 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 936.552217] env[62569]: WARNING nova.compute.manager [req-a3635fa4-39f1-4d1d-accf-a5e5a1131d97 req-2570ef26-7e6a-4886-afb2-0b1ea9c9b6f8 service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Received unexpected event network-vif-plugged-6ca4b28a-1255-4f88-a7b6-54b0378768b1 for instance with vm_state building and task_state spawning. [ 936.560522] env[62569]: DEBUG nova.compute.provider_tree [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.575844] env[62569]: DEBUG oslo_vmware.api [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250419, 'name': PowerOnVM_Task, 'duration_secs': 0.588437} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.576272] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 936.576481] env[62569]: INFO nova.compute.manager [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Took 8.78 seconds to spawn the instance on the hypervisor. [ 936.576665] env[62569]: DEBUG nova.compute.manager [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 936.577407] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c23f1f-207a-45f0-aacf-81dd7f91c489 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.637727] env[62569]: DEBUG nova.network.neutron [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Successfully updated port: 6ca4b28a-1255-4f88-a7b6-54b0378768b1 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 936.697390] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250420, 'name': CreateVM_Task, 'duration_secs': 0.361615} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.697524] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 936.698182] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.698361] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.698752] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 936.699026] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5d590ec-23fd-4c38-8772-e0d3af45ab5e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.703749] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 936.703749] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d527ac-40e3-7bf9-8832-ade55f2ac8bc" [ 936.703749] env[62569]: _type = "Task" [ 936.703749] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.712278] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d527ac-40e3-7bf9-8832-ade55f2ac8bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.817989] env[62569]: DEBUG oslo_concurrency.lockutils [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.063119] env[62569]: DEBUG nova.scheduler.client.report [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 937.094034] env[62569]: INFO nova.compute.manager [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Took 25.03 seconds to build instance. [ 937.141056] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.141056] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.141056] env[62569]: DEBUG nova.network.neutron [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 937.217666] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d527ac-40e3-7bf9-8832-ade55f2ac8bc, 'name': SearchDatastore_Task, 'duration_secs': 0.013856} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.218035] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.218298] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 937.218546] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.218709] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.218887] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 937.219208] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-727778c5-ea25-4fe2-9063-345e4400061c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.228055] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 937.228247] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 937.228958] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f178384-07c5-458f-a78b-8e6c33988edb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.233903] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 937.233903] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5282513b-49e6-1c05-2bf4-9393d4902145" [ 937.233903] env[62569]: _type = "Task" [ 937.233903] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.241088] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5282513b-49e6-1c05-2bf4-9393d4902145, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.568085] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.832s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.570531] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.296s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.570811] env[62569]: DEBUG nova.objects.instance [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Lazy-loading 'resources' on Instance uuid f751dfdb-e950-4821-b70d-4db78e03b9b7 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.594500] env[62569]: INFO nova.scheduler.client.report [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Deleted allocations for instance 6ccb78f3-d7f9-4090-9641-4633904efab8 [ 937.599328] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f67af71-e5e2-43a3-83f2-a4c3e1237111 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "af26f440-3515-4be7-9a03-8a0404c152d9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.551s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.689731] env[62569]: DEBUG nova.network.neutron [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 937.744337] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5282513b-49e6-1c05-2bf4-9393d4902145, 'name': SearchDatastore_Task, 'duration_secs': 0.010814} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.745129] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37129ec2-0044-4f48-91f5-ff74ab966c19 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.753018] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 937.753018] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529578dd-78a4-06bc-14b3-29535ff3198c" [ 937.753018] env[62569]: _type = "Task" [ 937.753018] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.760681] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529578dd-78a4-06bc-14b3-29535ff3198c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.886224] env[62569]: DEBUG nova.network.neutron [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Updating instance_info_cache with network_info: [{"id": "6ca4b28a-1255-4f88-a7b6-54b0378768b1", "address": "fa:16:3e:03:e7:ad", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ca4b28a-12", "ovs_interfaceid": "6ca4b28a-1255-4f88-a7b6-54b0378768b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.105595] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20a70c76-0fdc-46f0-9ed0-c8c520304254 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "6ccb78f3-d7f9-4090-9641-4633904efab8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.055s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.262652] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529578dd-78a4-06bc-14b3-29535ff3198c, 'name': SearchDatastore_Task, 'duration_secs': 0.014463} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.265134] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.265295] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] d1fdff70-0401-49ca-bbbb-ef5eda266c9c/d1fdff70-0401-49ca-bbbb-ef5eda266c9c.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 938.265896] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b7a4cb16-0b2a-45d9-87a8-7fb33814556c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.272837] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 938.272837] env[62569]: value = "task-1250421" [ 938.272837] env[62569]: _type = "Task" [ 938.272837] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.277216] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3a13d1f-142d-4d7b-946b-b79b0e052f69 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.284129] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250421, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.286731] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3fbaea7-a2c0-4767-9b3b-5c846f64fdb1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.318265] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aca4790a-766a-451d-8b01-698a6380eb7e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.325541] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749b3ab8-43e8-47dc-b8a9-28eff4f19ce1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.338820] env[62569]: DEBUG nova.compute.provider_tree [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.389585] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.389931] env[62569]: DEBUG nova.compute.manager [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Instance network_info: |[{"id": "6ca4b28a-1255-4f88-a7b6-54b0378768b1", "address": "fa:16:3e:03:e7:ad", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ca4b28a-12", "ovs_interfaceid": "6ca4b28a-1255-4f88-a7b6-54b0378768b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 938.390509] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:e7:ad', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '099fe970-c61f-4480-bed4-ae4f485fd82a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6ca4b28a-1255-4f88-a7b6-54b0378768b1', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 938.398368] env[62569]: DEBUG oslo.service.loopingcall [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.398593] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 938.398815] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7f97f9d5-1290-4ac4-98e4-268bc304b9a6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.418173] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 938.418173] env[62569]: value = "task-1250422" [ 938.418173] env[62569]: _type = "Task" [ 938.418173] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.425921] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250422, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.520332] env[62569]: DEBUG oslo_concurrency.lockutils [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "af26f440-3515-4be7-9a03-8a0404c152d9" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.520657] env[62569]: DEBUG oslo_concurrency.lockutils [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "af26f440-3515-4be7-9a03-8a0404c152d9" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.520808] env[62569]: INFO nova.compute.manager [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Shelving [ 938.578492] env[62569]: DEBUG nova.compute.manager [req-f7bc22bc-29c3-4dc8-ab8f-959707b20811 req-e177b97c-5a5a-4d77-bcc6-07bd6eb86f6f service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Received event network-changed-6ca4b28a-1255-4f88-a7b6-54b0378768b1 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 938.578774] env[62569]: DEBUG nova.compute.manager [req-f7bc22bc-29c3-4dc8-ab8f-959707b20811 req-e177b97c-5a5a-4d77-bcc6-07bd6eb86f6f service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Refreshing instance network info cache due to event network-changed-6ca4b28a-1255-4f88-a7b6-54b0378768b1. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 938.579046] env[62569]: DEBUG oslo_concurrency.lockutils [req-f7bc22bc-29c3-4dc8-ab8f-959707b20811 req-e177b97c-5a5a-4d77-bcc6-07bd6eb86f6f service nova] Acquiring lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.579277] env[62569]: DEBUG oslo_concurrency.lockutils [req-f7bc22bc-29c3-4dc8-ab8f-959707b20811 req-e177b97c-5a5a-4d77-bcc6-07bd6eb86f6f service nova] Acquired lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.579560] env[62569]: DEBUG nova.network.neutron [req-f7bc22bc-29c3-4dc8-ab8f-959707b20811 req-e177b97c-5a5a-4d77-bcc6-07bd6eb86f6f service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Refreshing network info cache for port 6ca4b28a-1255-4f88-a7b6-54b0378768b1 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 938.784819] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250421, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.842471] env[62569]: DEBUG nova.scheduler.client.report [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 938.930800] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250422, 'name': CreateVM_Task, 'duration_secs': 0.449311} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.931346] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 938.931765] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.931939] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.932367] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 938.932672] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78e37c37-f487-4141-9750-f564ed106298 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.937399] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 938.937399] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d9a3fd-6040-47f4-0eb4-9b41d30a68a8" [ 938.937399] env[62569]: _type = "Task" [ 938.937399] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.945958] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d9a3fd-6040-47f4-0eb4-9b41d30a68a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.286554] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250421, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.682367} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.286873] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] d1fdff70-0401-49ca-bbbb-ef5eda266c9c/d1fdff70-0401-49ca-bbbb-ef5eda266c9c.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 939.287115] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 939.287666] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5ef9e4c1-7363-42e3-8cca-fafeb295973b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.294660] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 939.294660] env[62569]: value = "task-1250423" [ 939.294660] env[62569]: _type = "Task" [ 939.294660] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.306271] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250423, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.347703] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.777s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.350421] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.818s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.350665] env[62569]: DEBUG nova.objects.instance [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lazy-loading 'resources' on Instance uuid 337d88a2-30b9-4846-929e-042bd7a64a65 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 939.373462] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "abf56021-ddf8-4677-9e90-46dc96e89133" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.373950] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "abf56021-ddf8-4677-9e90-46dc96e89133" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.377785] env[62569]: INFO nova.scheduler.client.report [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Deleted allocations for instance f751dfdb-e950-4821-b70d-4db78e03b9b7 [ 939.450117] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d9a3fd-6040-47f4-0eb4-9b41d30a68a8, 'name': SearchDatastore_Task, 'duration_secs': 0.009991} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.450117] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.450117] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 939.450318] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.450365] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.450962] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 939.451114] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c2ab7279-80d3-40bc-9469-59ca4023ccc8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.462548] env[62569]: DEBUG nova.network.neutron [req-f7bc22bc-29c3-4dc8-ab8f-959707b20811 req-e177b97c-5a5a-4d77-bcc6-07bd6eb86f6f service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Updated VIF entry in instance network info cache for port 6ca4b28a-1255-4f88-a7b6-54b0378768b1. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 939.463177] env[62569]: DEBUG nova.network.neutron [req-f7bc22bc-29c3-4dc8-ab8f-959707b20811 req-e177b97c-5a5a-4d77-bcc6-07bd6eb86f6f service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Updating instance_info_cache with network_info: [{"id": "6ca4b28a-1255-4f88-a7b6-54b0378768b1", "address": "fa:16:3e:03:e7:ad", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ca4b28a-12", "ovs_interfaceid": "6ca4b28a-1255-4f88-a7b6-54b0378768b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.465101] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 939.465285] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 939.466232] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c74474c-457e-4a38-9a79-1090611ab493 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.472275] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 939.472275] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5220990b-cfaa-a661-d746-c0d57e8ba7bc" [ 939.472275] env[62569]: _type = "Task" [ 939.472275] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.481764] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5220990b-cfaa-a661-d746-c0d57e8ba7bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.532561] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 939.532911] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-60632497-c314-4546-aec1-375b7a1784bf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.540423] env[62569]: DEBUG oslo_vmware.api [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 939.540423] env[62569]: value = "task-1250424" [ 939.540423] env[62569]: _type = "Task" [ 939.540423] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.549333] env[62569]: DEBUG oslo_vmware.api [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250424, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.805045] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250423, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062934} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.805249] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 939.806064] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5c787ca-1b45-41af-b2c1-1ba8d13cf40a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.830719] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] d1fdff70-0401-49ca-bbbb-ef5eda266c9c/d1fdff70-0401-49ca-bbbb-ef5eda266c9c.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 939.831071] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1d764155-8bda-4fb2-aa37-6a981288b0e7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.854181] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 939.854181] env[62569]: value = "task-1250425" [ 939.854181] env[62569]: _type = "Task" [ 939.854181] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.865667] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250425, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.877926] env[62569]: DEBUG nova.compute.manager [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 939.888244] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3484b8a9-701d-4d43-869d-3cb7c93b6d34 tempest-InstanceActionsNegativeTestJSON-483609019 tempest-InstanceActionsNegativeTestJSON-483609019-project-member] Lock "f751dfdb-e950-4821-b70d-4db78e03b9b7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.101s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.967124] env[62569]: DEBUG oslo_concurrency.lockutils [req-f7bc22bc-29c3-4dc8-ab8f-959707b20811 req-e177b97c-5a5a-4d77-bcc6-07bd6eb86f6f service nova] Releasing lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.985714] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5220990b-cfaa-a661-d746-c0d57e8ba7bc, 'name': SearchDatastore_Task, 'duration_secs': 0.027629} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.989355] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c6be1e3-7ecf-4049-8e9f-ec9a4e3c635b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.995280] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 939.995280] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ac0a9c-0067-e3ed-c240-baa5e8be05d0" [ 939.995280] env[62569]: _type = "Task" [ 939.995280] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.005495] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ac0a9c-0067-e3ed-c240-baa5e8be05d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.050264] env[62569]: DEBUG oslo_vmware.api [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250424, 'name': PowerOffVM_Task, 'duration_secs': 0.361048} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.052471] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 940.053808] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc7985ad-e1f4-4d03-92a2-61363d878082 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.074654] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac520b6-98e2-4b29-a2ae-69c8066a20af {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.099674] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ede5908a-4bfd-42d1-aa69-457d6b277111 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.107284] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c4ab1b-e47c-4026-8d33-bef70af9fdcd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.137805] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-594a3df4-c581-4e90-9ce5-7242ff6f3e29 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.145894] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4164958f-2ea4-4ffb-b250-a0cea43204b8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.160781] env[62569]: DEBUG nova.compute.provider_tree [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 940.363431] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250425, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.395947] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.508063] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ac0a9c-0067-e3ed-c240-baa5e8be05d0, 'name': SearchDatastore_Task, 'duration_secs': 0.011986} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.508063] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.508063] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e/93eb1a09-269a-44c8-8f9d-3a2b7e0d715e.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 940.508063] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1b84f943-69a8-43b4-b820-92c2004eda5e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.513978] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 940.513978] env[62569]: value = "task-1250426" [ 940.513978] env[62569]: _type = "Task" [ 940.513978] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.522236] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250426, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.588170] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Creating Snapshot of the VM instance {{(pid=62569) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 940.588170] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-2351bcb0-af64-4534-b529-1351dc2876f5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.595018] env[62569]: DEBUG oslo_vmware.api [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 940.595018] env[62569]: value = "task-1250427" [ 940.595018] env[62569]: _type = "Task" [ 940.595018] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.602392] env[62569]: DEBUG oslo_vmware.api [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250427, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.663837] env[62569]: DEBUG nova.scheduler.client.report [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 940.868028] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250425, 'name': ReconfigVM_Task, 'duration_secs': 0.540687} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.868028] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Reconfigured VM instance instance-0000005d to attach disk [datastore2] d1fdff70-0401-49ca-bbbb-ef5eda266c9c/d1fdff70-0401-49ca-bbbb-ef5eda266c9c.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 940.868028] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-33ee71b9-991e-4c1d-956a-0add84275dff {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.876227] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 940.876227] env[62569]: value = "task-1250428" [ 940.876227] env[62569]: _type = "Task" [ 940.876227] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.885316] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250428, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.025029] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250426, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.469636} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.025029] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e/93eb1a09-269a-44c8-8f9d-3a2b7e0d715e.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 941.025029] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 941.028018] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fb6d9f29-1ff8-4727-919f-3e6c51c3d10b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.032791] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 941.032791] env[62569]: value = "task-1250429" [ 941.032791] env[62569]: _type = "Task" [ 941.032791] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.041779] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250429, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.104143] env[62569]: DEBUG oslo_vmware.api [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250427, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.173152] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.819s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.173152] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.473s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.175271] env[62569]: DEBUG nova.objects.instance [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lazy-loading 'pci_requests' on Instance uuid b6e5eefc-8c06-445b-a3af-9404578b6179 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 941.203546] env[62569]: INFO nova.scheduler.client.report [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Deleted allocations for instance 337d88a2-30b9-4846-929e-042bd7a64a65 [ 941.385027] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250428, 'name': Rename_Task, 'duration_secs': 0.156806} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.385189] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 941.385494] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ed02157-56d7-46dc-91e5-5ec8e7213b04 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.392191] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 941.392191] env[62569]: value = "task-1250430" [ 941.392191] env[62569]: _type = "Task" [ 941.392191] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.399638] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250430, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.544419] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250429, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06359} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.544710] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 941.545526] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d3617c0-4b72-4298-92b6-6f388dd1d7cf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.569022] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e/93eb1a09-269a-44c8-8f9d-3a2b7e0d715e.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 941.569352] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1af00393-74ea-46de-91fb-e14401b63ee7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.588979] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 941.588979] env[62569]: value = "task-1250431" [ 941.588979] env[62569]: _type = "Task" [ 941.588979] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.597175] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250431, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.607051] env[62569]: DEBUG oslo_vmware.api [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250427, 'name': CreateSnapshot_Task, 'duration_secs': 0.968585} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.607439] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Created Snapshot of the VM instance {{(pid=62569) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 941.608190] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-320c579b-7b38-4e41-b312-5f0832d3cbee {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.680512] env[62569]: DEBUG nova.objects.instance [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lazy-loading 'numa_topology' on Instance uuid b6e5eefc-8c06-445b-a3af-9404578b6179 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 941.711570] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de0e2c28-339d-4d4b-bac3-fd45e1b024b5 tempest-AttachVolumeShelveTestJSON-1327975016 tempest-AttachVolumeShelveTestJSON-1327975016-project-member] Lock "337d88a2-30b9-4846-929e-042bd7a64a65" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.129s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.906648] env[62569]: DEBUG oslo_vmware.api [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250430, 'name': PowerOnVM_Task, 'duration_secs': 0.46533} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.906959] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 941.907258] env[62569]: INFO nova.compute.manager [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Took 9.39 seconds to spawn the instance on the hypervisor. [ 941.907462] env[62569]: DEBUG nova.compute.manager [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 941.908269] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-973e6f61-fa0c-4673-b6da-7690f21eafed {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.099142] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250431, 'name': ReconfigVM_Task, 'duration_secs': 0.312903} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.099496] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e/93eb1a09-269a-44c8-8f9d-3a2b7e0d715e.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 942.100238] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-83463cc7-e777-427f-b440-64747ecae069 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.107286] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 942.107286] env[62569]: value = "task-1250432" [ 942.107286] env[62569]: _type = "Task" [ 942.107286] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.116035] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250432, 'name': Rename_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.125758] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Creating linked-clone VM from snapshot {{(pid=62569) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 942.125953] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-39239779-3eda-40f9-bca5-cd114bab6bf6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.133815] env[62569]: DEBUG oslo_vmware.api [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 942.133815] env[62569]: value = "task-1250433" [ 942.133815] env[62569]: _type = "Task" [ 942.133815] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.142171] env[62569]: DEBUG oslo_vmware.api [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250433, 'name': CloneVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.184100] env[62569]: INFO nova.compute.claims [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 942.436150] env[62569]: INFO nova.compute.manager [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Took 22.23 seconds to build instance. [ 942.618339] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250432, 'name': Rename_Task, 'duration_secs': 0.310507} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.618648] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 942.618902] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b60377b3-8249-4775-9705-cc23d562cf18 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.624486] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 942.624486] env[62569]: value = "task-1250434" [ 942.624486] env[62569]: _type = "Task" [ 942.624486] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.632952] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250434, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.641293] env[62569]: DEBUG oslo_vmware.api [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250433, 'name': CloneVM_Task} progress is 94%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.938704] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2a33a611-98d8-4089-80c7-66d987731243 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "d1fdff70-0401-49ca-bbbb-ef5eda266c9c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.739s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.135487] env[62569]: DEBUG oslo_vmware.api [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250434, 'name': PowerOnVM_Task, 'duration_secs': 0.431315} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.138569] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 943.138794] env[62569]: INFO nova.compute.manager [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Took 7.29 seconds to spawn the instance on the hypervisor. [ 943.138988] env[62569]: DEBUG nova.compute.manager [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 943.139828] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4ec2641-ce83-4271-b8a4-7bb2de6dc121 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.147262] env[62569]: DEBUG oslo_vmware.api [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250433, 'name': CloneVM_Task} progress is 94%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.388258] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c9db05-98e1-4a57-b833-0a77b1dd2e1d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.396644] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b47dcf53-1f24-43c0-93bb-561e3f58ce6f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.427858] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c38be77-d756-4d0b-9409-20438deb5e3e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.436238] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d4d24b0-d895-4a39-9e19-5b561e66ebfd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.450871] env[62569]: DEBUG nova.compute.provider_tree [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.650067] env[62569]: DEBUG oslo_vmware.api [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250433, 'name': CloneVM_Task, 'duration_secs': 1.351109} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.650067] env[62569]: INFO nova.virt.vmwareapi.vmops [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Created linked-clone VM from snapshot [ 943.650365] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-389c5d5d-0a84-4394-9582-a549225f4f4a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.670157] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Uploading image deff6cce-4781-495b-a37b-642db84153cd {{(pid=62569) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 943.670879] env[62569]: INFO nova.compute.manager [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Took 22.08 seconds to build instance. [ 943.696143] env[62569]: DEBUG oslo_vmware.rw_handles [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 943.696143] env[62569]: value = "vm-269481" [ 943.696143] env[62569]: _type = "VirtualMachine" [ 943.696143] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 943.696143] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-45fda985-865b-4771-81ce-434dcddc38c3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.703796] env[62569]: DEBUG oslo_vmware.rw_handles [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lease: (returnval){ [ 943.703796] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ab94f1-cf7d-a3b8-4ac4-ba79f551a9e4" [ 943.703796] env[62569]: _type = "HttpNfcLease" [ 943.703796] env[62569]: } obtained for exporting VM: (result){ [ 943.703796] env[62569]: value = "vm-269481" [ 943.703796] env[62569]: _type = "VirtualMachine" [ 943.703796] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 943.704457] env[62569]: DEBUG oslo_vmware.api [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the lease: (returnval){ [ 943.704457] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ab94f1-cf7d-a3b8-4ac4-ba79f551a9e4" [ 943.704457] env[62569]: _type = "HttpNfcLease" [ 943.704457] env[62569]: } to be ready. {{(pid=62569) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 943.711629] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 943.711629] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ab94f1-cf7d-a3b8-4ac4-ba79f551a9e4" [ 943.711629] env[62569]: _type = "HttpNfcLease" [ 943.711629] env[62569]: } is initializing. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 943.770904] env[62569]: DEBUG nova.compute.manager [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 943.770904] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8a0f376-1ef2-49e7-beb1-9b369fff7577 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.954642] env[62569]: DEBUG nova.scheduler.client.report [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 944.175413] env[62569]: DEBUG oslo_concurrency.lockutils [None req-68572c29-fc2b-4442-a122-70480674bbfa tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.589s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.214151] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 944.214151] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ab94f1-cf7d-a3b8-4ac4-ba79f551a9e4" [ 944.214151] env[62569]: _type = "HttpNfcLease" [ 944.214151] env[62569]: } is ready. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 944.214458] env[62569]: DEBUG oslo_vmware.rw_handles [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 944.214458] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ab94f1-cf7d-a3b8-4ac4-ba79f551a9e4" [ 944.214458] env[62569]: _type = "HttpNfcLease" [ 944.214458] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 944.215311] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d6de873-1ffd-44c6-b4f2-65c384e38447 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.223181] env[62569]: DEBUG oslo_vmware.rw_handles [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fbe02f-5904-61cd-6beb-d306008f74fd/disk-0.vmdk from lease info. {{(pid=62569) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 944.223371] env[62569]: DEBUG oslo_vmware.rw_handles [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fbe02f-5904-61cd-6beb-d306008f74fd/disk-0.vmdk for reading. {{(pid=62569) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 944.300486] env[62569]: INFO nova.compute.manager [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] instance snapshotting [ 944.304286] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deaaa0c4-d931-4043-a7a1-4dd41a4e4bbd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.326124] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cf637fc-6ce6-4d0d-a588-f603162dc163 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.341558] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-577e3459-9d5f-4fb7-8378-609fade591b8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.460364] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.288s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.464023] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 14.293s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.551275] env[62569]: INFO nova.network.neutron [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Updating port cfddbae7-9ab0-4bc6-aea1-46a0cde743fc with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 944.839891] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Creating Snapshot of the VM instance {{(pid=62569) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 944.843645] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c3930519-5767-4d34-a3bd-5ac29e0bc5a4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.848640] env[62569]: DEBUG oslo_vmware.api [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 944.848640] env[62569]: value = "task-1250436" [ 944.848640] env[62569]: _type = "Task" [ 944.848640] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.867260] env[62569]: DEBUG oslo_vmware.api [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250436, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.968759] env[62569]: INFO nova.compute.claims [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 945.372532] env[62569]: DEBUG oslo_vmware.api [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250436, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.475993] env[62569]: INFO nova.compute.resource_tracker [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Updating resource usage from migration 54fd4982-eb4e-4229-a3e0-21648ce18eb7 [ 945.537235] env[62569]: DEBUG nova.compute.manager [req-326ce1e6-7b5a-45c1-b62e-02d7069126e8 req-80a9ece6-bf43-462a-8be6-f455cffd2d6d service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Received event network-changed-91986023-125e-442b-b105-102630ec6036 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 945.537235] env[62569]: DEBUG nova.compute.manager [req-326ce1e6-7b5a-45c1-b62e-02d7069126e8 req-80a9ece6-bf43-462a-8be6-f455cffd2d6d service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Refreshing instance network info cache due to event network-changed-91986023-125e-442b-b105-102630ec6036. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 945.537235] env[62569]: DEBUG oslo_concurrency.lockutils [req-326ce1e6-7b5a-45c1-b62e-02d7069126e8 req-80a9ece6-bf43-462a-8be6-f455cffd2d6d service nova] Acquiring lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.537235] env[62569]: DEBUG oslo_concurrency.lockutils [req-326ce1e6-7b5a-45c1-b62e-02d7069126e8 req-80a9ece6-bf43-462a-8be6-f455cffd2d6d service nova] Acquired lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.537235] env[62569]: DEBUG nova.network.neutron [req-326ce1e6-7b5a-45c1-b62e-02d7069126e8 req-80a9ece6-bf43-462a-8be6-f455cffd2d6d service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Refreshing network info cache for port 91986023-125e-442b-b105-102630ec6036 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 945.742185] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fef5c755-5588-408b-9f59-5f5e28724eea {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.754820] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b66da33d-7a4b-4bc3-93e4-a54a6c6660a5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.789130] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e949941c-7c8e-45f5-add8-83e208c6ba94 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.798089] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-699e8163-a508-4566-ba8c-16cf6d968f7c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.816317] env[62569]: DEBUG nova.compute.provider_tree [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.862508] env[62569]: DEBUG oslo_vmware.api [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250436, 'name': CreateSnapshot_Task, 'duration_secs': 0.662307} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.862835] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Created Snapshot of the VM instance {{(pid=62569) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 945.863657] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53df2a4f-24fe-4b50-b49f-b83e840c0a01 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.274123] env[62569]: DEBUG nova.network.neutron [req-326ce1e6-7b5a-45c1-b62e-02d7069126e8 req-80a9ece6-bf43-462a-8be6-f455cffd2d6d service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Updated VIF entry in instance network info cache for port 91986023-125e-442b-b105-102630ec6036. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 946.274123] env[62569]: DEBUG nova.network.neutron [req-326ce1e6-7b5a-45c1-b62e-02d7069126e8 req-80a9ece6-bf43-462a-8be6-f455cffd2d6d service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Updating instance_info_cache with network_info: [{"id": "91986023-125e-442b-b105-102630ec6036", "address": "fa:16:3e:18:83:9b", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91986023-12", "ovs_interfaceid": "91986023-125e-442b-b105-102630ec6036", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.303355] env[62569]: DEBUG nova.compute.manager [req-9caa337c-aed6-4d56-bd42-55f01dfde79d req-d8680a14-3280-45d5-9ef7-bbcde4fbb323 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Received event network-vif-plugged-cfddbae7-9ab0-4bc6-aea1-46a0cde743fc {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 946.303355] env[62569]: DEBUG oslo_concurrency.lockutils [req-9caa337c-aed6-4d56-bd42-55f01dfde79d req-d8680a14-3280-45d5-9ef7-bbcde4fbb323 service nova] Acquiring lock "b6e5eefc-8c06-445b-a3af-9404578b6179-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.303355] env[62569]: DEBUG oslo_concurrency.lockutils [req-9caa337c-aed6-4d56-bd42-55f01dfde79d req-d8680a14-3280-45d5-9ef7-bbcde4fbb323 service nova] Lock "b6e5eefc-8c06-445b-a3af-9404578b6179-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.303553] env[62569]: DEBUG oslo_concurrency.lockutils [req-9caa337c-aed6-4d56-bd42-55f01dfde79d req-d8680a14-3280-45d5-9ef7-bbcde4fbb323 service nova] Lock "b6e5eefc-8c06-445b-a3af-9404578b6179-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.304128] env[62569]: DEBUG nova.compute.manager [req-9caa337c-aed6-4d56-bd42-55f01dfde79d req-d8680a14-3280-45d5-9ef7-bbcde4fbb323 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] No waiting events found dispatching network-vif-plugged-cfddbae7-9ab0-4bc6-aea1-46a0cde743fc {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 946.304128] env[62569]: WARNING nova.compute.manager [req-9caa337c-aed6-4d56-bd42-55f01dfde79d req-d8680a14-3280-45d5-9ef7-bbcde4fbb323 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Received unexpected event network-vif-plugged-cfddbae7-9ab0-4bc6-aea1-46a0cde743fc for instance with vm_state shelved_offloaded and task_state spawning. [ 946.319044] env[62569]: DEBUG nova.scheduler.client.report [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 946.381646] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Creating linked-clone VM from snapshot {{(pid=62569) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 946.382081] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1d0747b1-b872-4139-ac7b-3bc3c8ea2e42 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.391548] env[62569]: DEBUG oslo_vmware.api [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 946.391548] env[62569]: value = "task-1250437" [ 946.391548] env[62569]: _type = "Task" [ 946.391548] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.407793] env[62569]: DEBUG oslo_vmware.api [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250437, 'name': CloneVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.777272] env[62569]: DEBUG oslo_concurrency.lockutils [req-326ce1e6-7b5a-45c1-b62e-02d7069126e8 req-80a9ece6-bf43-462a-8be6-f455cffd2d6d service nova] Releasing lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.824014] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.361s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.824286] env[62569]: INFO nova.compute.manager [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Migrating [ 946.831832] env[62569]: DEBUG oslo_concurrency.lockutils [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 11.937s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.888534] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.888741] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquired lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.888957] env[62569]: DEBUG nova.network.neutron [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 946.908027] env[62569]: DEBUG oslo_vmware.api [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250437, 'name': CloneVM_Task} progress is 94%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.343781] env[62569]: INFO nova.compute.claims [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 947.347905] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.348131] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.348331] env[62569]: DEBUG nova.network.neutron [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 947.412817] env[62569]: DEBUG oslo_vmware.api [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250437, 'name': CloneVM_Task} progress is 94%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.557815] env[62569]: DEBUG nova.compute.manager [req-84c6a834-1a02-43a1-a4d2-e6dc2c4e8c05 req-2baff221-3e48-4bd2-8bd3-1110e3249e91 service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Received event network-changed-6ca4b28a-1255-4f88-a7b6-54b0378768b1 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 947.557923] env[62569]: DEBUG nova.compute.manager [req-84c6a834-1a02-43a1-a4d2-e6dc2c4e8c05 req-2baff221-3e48-4bd2-8bd3-1110e3249e91 service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Refreshing instance network info cache due to event network-changed-6ca4b28a-1255-4f88-a7b6-54b0378768b1. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 947.558157] env[62569]: DEBUG oslo_concurrency.lockutils [req-84c6a834-1a02-43a1-a4d2-e6dc2c4e8c05 req-2baff221-3e48-4bd2-8bd3-1110e3249e91 service nova] Acquiring lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.558323] env[62569]: DEBUG oslo_concurrency.lockutils [req-84c6a834-1a02-43a1-a4d2-e6dc2c4e8c05 req-2baff221-3e48-4bd2-8bd3-1110e3249e91 service nova] Acquired lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.558492] env[62569]: DEBUG nova.network.neutron [req-84c6a834-1a02-43a1-a4d2-e6dc2c4e8c05 req-2baff221-3e48-4bd2-8bd3-1110e3249e91 service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Refreshing network info cache for port 6ca4b28a-1255-4f88-a7b6-54b0378768b1 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 947.755809] env[62569]: DEBUG nova.network.neutron [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Updating instance_info_cache with network_info: [{"id": "cfddbae7-9ab0-4bc6-aea1-46a0cde743fc", "address": "fa:16:3e:b7:d6:80", "network": {"id": "406dd658-8168-4e73-9b97-861929bfaa2e", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-693938985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de8605118d744a93bce54e897aa849b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfddbae7-9a", "ovs_interfaceid": "cfddbae7-9ab0-4bc6-aea1-46a0cde743fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.854137] env[62569]: INFO nova.compute.resource_tracker [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Updating resource usage from migration 5b4b811b-c7c1-4e18-b28c-14df9b5a62ed [ 947.915926] env[62569]: DEBUG oslo_vmware.api [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250437, 'name': CloneVM_Task} progress is 95%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.111774] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d3f905-05de-4713-b078-5f57ebe3fec2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.120226] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1de8f0a4-97a3-4e47-8533-229ca084b784 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.155924] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa5cd0da-ae12-41df-833d-beb7ef707b64 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.164090] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-538c2488-ae78-40c7-97e8-71e66280a6a1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.181703] env[62569]: DEBUG nova.compute.provider_tree [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.221332] env[62569]: DEBUG nova.network.neutron [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Updating instance_info_cache with network_info: [{"id": "90b3aaef-2e72-486f-be8d-a4202a6cb0c6", "address": "fa:16:3e:0c:8f:dd", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90b3aaef-2e", "ovs_interfaceid": "90b3aaef-2e72-486f-be8d-a4202a6cb0c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.258456] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Releasing lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.290452] env[62569]: DEBUG nova.virt.hardware [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='d7604b9a3b04630f4fe0e0e1576a86cc',container_format='bare',created_at=2024-10-10T15:38:23Z,direct_url=,disk_format='vmdk',id=87099cdd-38b2-4615-944d-a91a20479717,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-103782614-shelved',owner='de8605118d744a93bce54e897aa849b5',properties=ImageMetaProps,protected=,size=31665664,status='active',tags=,updated_at=2024-10-10T15:38:38Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 948.290993] env[62569]: DEBUG nova.virt.hardware [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 948.290993] env[62569]: DEBUG nova.virt.hardware [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 948.291197] env[62569]: DEBUG nova.virt.hardware [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 948.291278] env[62569]: DEBUG nova.virt.hardware [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 948.291452] env[62569]: DEBUG nova.virt.hardware [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 948.291693] env[62569]: DEBUG nova.virt.hardware [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 948.292844] env[62569]: DEBUG nova.virt.hardware [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 948.292844] env[62569]: DEBUG nova.virt.hardware [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 948.292844] env[62569]: DEBUG nova.virt.hardware [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 948.292844] env[62569]: DEBUG nova.virt.hardware [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 948.293481] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59925ad7-f51b-40be-a857-642b692eb0f4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.304425] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6d69374-2820-4450-8797-c1aaf1374c6a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.319204] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b7:d6:80', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '257e5ea7-8b80-4301-9900-a754f1fe2031', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cfddbae7-9ab0-4bc6-aea1-46a0cde743fc', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 948.327208] env[62569]: DEBUG oslo.service.loopingcall [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 948.328295] env[62569]: DEBUG nova.network.neutron [req-84c6a834-1a02-43a1-a4d2-e6dc2c4e8c05 req-2baff221-3e48-4bd2-8bd3-1110e3249e91 service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Updated VIF entry in instance network info cache for port 6ca4b28a-1255-4f88-a7b6-54b0378768b1. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 948.328685] env[62569]: DEBUG nova.network.neutron [req-84c6a834-1a02-43a1-a4d2-e6dc2c4e8c05 req-2baff221-3e48-4bd2-8bd3-1110e3249e91 service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Updating instance_info_cache with network_info: [{"id": "6ca4b28a-1255-4f88-a7b6-54b0378768b1", "address": "fa:16:3e:03:e7:ad", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ca4b28a-12", "ovs_interfaceid": "6ca4b28a-1255-4f88-a7b6-54b0378768b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.329958] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 948.330421] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4500f77a-230c-4f75-801b-1a2dc0a52f7d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.349193] env[62569]: DEBUG nova.compute.manager [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Received event network-changed-cfddbae7-9ab0-4bc6-aea1-46a0cde743fc {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 948.349471] env[62569]: DEBUG nova.compute.manager [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Refreshing instance network info cache due to event network-changed-cfddbae7-9ab0-4bc6-aea1-46a0cde743fc. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 948.349705] env[62569]: DEBUG oslo_concurrency.lockutils [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] Acquiring lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.349853] env[62569]: DEBUG oslo_concurrency.lockutils [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] Acquired lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.350032] env[62569]: DEBUG nova.network.neutron [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Refreshing network info cache for port cfddbae7-9ab0-4bc6-aea1-46a0cde743fc {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 948.354246] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 948.354246] env[62569]: value = "task-1250438" [ 948.354246] env[62569]: _type = "Task" [ 948.354246] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.362794] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250438, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.406583] env[62569]: DEBUG oslo_vmware.api [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250437, 'name': CloneVM_Task, 'duration_secs': 1.658822} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.406855] env[62569]: INFO nova.virt.vmwareapi.vmops [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Created linked-clone VM from snapshot [ 948.407797] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-033a3968-db67-444c-ab6b-8616ee36add0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.415488] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Uploading image dcad6631-f683-4ed5-9957-aac1d800c83e {{(pid=62569) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 948.426456] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Destroying the VM {{(pid=62569) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 948.426552] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ccbffcf8-d9ff-4ac9-8a79-62063583b9d7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.433713] env[62569]: DEBUG oslo_vmware.api [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 948.433713] env[62569]: value = "task-1250439" [ 948.433713] env[62569]: _type = "Task" [ 948.433713] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.442345] env[62569]: DEBUG oslo_vmware.api [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250439, 'name': Destroy_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.686796] env[62569]: DEBUG nova.scheduler.client.report [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 948.724181] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Releasing lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.831802] env[62569]: DEBUG oslo_concurrency.lockutils [req-84c6a834-1a02-43a1-a4d2-e6dc2c4e8c05 req-2baff221-3e48-4bd2-8bd3-1110e3249e91 service nova] Releasing lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.865775] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250438, 'name': CreateVM_Task, 'duration_secs': 0.493042} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.865775] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 948.866869] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/87099cdd-38b2-4615-944d-a91a20479717" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.866869] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquired lock "[datastore1] devstack-image-cache_base/87099cdd-38b2-4615-944d-a91a20479717" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.867060] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/87099cdd-38b2-4615-944d-a91a20479717" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 948.867252] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cab1e291-623d-44f9-a04a-10cfca1a2154 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.872315] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 948.872315] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52861623-5ba1-8652-9175-8846b1ea3f83" [ 948.872315] env[62569]: _type = "Task" [ 948.872315] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.881188] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52861623-5ba1-8652-9175-8846b1ea3f83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.947705] env[62569]: DEBUG oslo_vmware.api [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250439, 'name': Destroy_Task} progress is 33%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.099297] env[62569]: DEBUG nova.network.neutron [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Updated VIF entry in instance network info cache for port cfddbae7-9ab0-4bc6-aea1-46a0cde743fc. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 949.100022] env[62569]: DEBUG nova.network.neutron [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Updating instance_info_cache with network_info: [{"id": "cfddbae7-9ab0-4bc6-aea1-46a0cde743fc", "address": "fa:16:3e:b7:d6:80", "network": {"id": "406dd658-8168-4e73-9b97-861929bfaa2e", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-693938985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de8605118d744a93bce54e897aa849b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfddbae7-9a", "ovs_interfaceid": "cfddbae7-9ab0-4bc6-aea1-46a0cde743fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.192567] env[62569]: DEBUG oslo_concurrency.lockutils [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.361s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.192817] env[62569]: INFO nova.compute.manager [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Migrating [ 949.199428] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 13.393s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.383080] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Releasing lock "[datastore1] devstack-image-cache_base/87099cdd-38b2-4615-944d-a91a20479717" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.383080] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Processing image 87099cdd-38b2-4615-944d-a91a20479717 {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 949.383313] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/87099cdd-38b2-4615-944d-a91a20479717/87099cdd-38b2-4615-944d-a91a20479717.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.383428] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquired lock "[datastore1] devstack-image-cache_base/87099cdd-38b2-4615-944d-a91a20479717/87099cdd-38b2-4615-944d-a91a20479717.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.383673] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 949.383980] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-759e7680-c658-4b7a-85af-8cae7e1aafe2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.392801] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 949.393041] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 949.393834] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4404bf9f-5445-45bd-8143-34f989b1992a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.399103] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 949.399103] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b7221a-de9e-7dbe-d88e-1492b866e6b6" [ 949.399103] env[62569]: _type = "Task" [ 949.399103] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.413770] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Preparing fetch location {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 949.414071] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Fetch image to [datastore1] OSTACK_IMG_6a50e58e-0561-481e-bf5f-89bb6d1df7d9/OSTACK_IMG_6a50e58e-0561-481e-bf5f-89bb6d1df7d9.vmdk {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 949.414266] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Downloading stream optimized image 87099cdd-38b2-4615-944d-a91a20479717 to [datastore1] OSTACK_IMG_6a50e58e-0561-481e-bf5f-89bb6d1df7d9/OSTACK_IMG_6a50e58e-0561-481e-bf5f-89bb6d1df7d9.vmdk on the data store datastore1 as vApp {{(pid=62569) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 949.414450] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Downloading image file data 87099cdd-38b2-4615-944d-a91a20479717 to the ESX as VM named 'OSTACK_IMG_6a50e58e-0561-481e-bf5f-89bb6d1df7d9' {{(pid=62569) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 949.461207] env[62569]: DEBUG oslo_vmware.api [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250439, 'name': Destroy_Task, 'duration_secs': 0.580188} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.461614] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Destroyed the VM [ 949.461850] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Deleting Snapshot of the VM instance {{(pid=62569) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 949.462069] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-79210d89-dae7-402a-bb1f-01b6ffd6b271 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.468372] env[62569]: DEBUG oslo_vmware.api [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 949.468372] env[62569]: value = "task-1250440" [ 949.468372] env[62569]: _type = "Task" [ 949.468372] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.478650] env[62569]: DEBUG oslo_vmware.api [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250440, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.493325] env[62569]: DEBUG oslo_vmware.rw_handles [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 949.493325] env[62569]: value = "resgroup-9" [ 949.493325] env[62569]: _type = "ResourcePool" [ 949.493325] env[62569]: }. {{(pid=62569) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 949.493706] env[62569]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-0b93cc13-9904-403b-a028-682539e37db1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.514786] env[62569]: DEBUG oslo_vmware.rw_handles [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lease: (returnval){ [ 949.514786] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52169769-4e82-75d3-1a54-e1c24183716e" [ 949.514786] env[62569]: _type = "HttpNfcLease" [ 949.514786] env[62569]: } obtained for vApp import into resource pool (val){ [ 949.514786] env[62569]: value = "resgroup-9" [ 949.514786] env[62569]: _type = "ResourcePool" [ 949.514786] env[62569]: }. {{(pid=62569) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 949.515370] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the lease: (returnval){ [ 949.515370] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52169769-4e82-75d3-1a54-e1c24183716e" [ 949.515370] env[62569]: _type = "HttpNfcLease" [ 949.515370] env[62569]: } to be ready. {{(pid=62569) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 949.521708] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 949.521708] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52169769-4e82-75d3-1a54-e1c24183716e" [ 949.521708] env[62569]: _type = "HttpNfcLease" [ 949.521708] env[62569]: } is initializing. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 949.603075] env[62569]: DEBUG oslo_concurrency.lockutils [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] Releasing lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.603415] env[62569]: DEBUG nova.compute.manager [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Received event network-changed-6ca4b28a-1255-4f88-a7b6-54b0378768b1 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 949.603597] env[62569]: DEBUG nova.compute.manager [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Refreshing instance network info cache due to event network-changed-6ca4b28a-1255-4f88-a7b6-54b0378768b1. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 949.603829] env[62569]: DEBUG oslo_concurrency.lockutils [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] Acquiring lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.603975] env[62569]: DEBUG oslo_concurrency.lockutils [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] Acquired lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.604161] env[62569]: DEBUG nova.network.neutron [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Refreshing network info cache for port 6ca4b28a-1255-4f88-a7b6-54b0378768b1 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 949.715341] env[62569]: DEBUG oslo_concurrency.lockutils [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "refresh_cache-09eeb937-aabc-418d-9fd3-3c7a72f4ef75" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.715563] env[62569]: DEBUG oslo_concurrency.lockutils [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired lock "refresh_cache-09eeb937-aabc-418d-9fd3-3c7a72f4ef75" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.715704] env[62569]: DEBUG nova.network.neutron [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 949.982711] env[62569]: DEBUG oslo_vmware.api [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250440, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.023023] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 950.023023] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52169769-4e82-75d3-1a54-e1c24183716e" [ 950.023023] env[62569]: _type = "HttpNfcLease" [ 950.023023] env[62569]: } is ready. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 950.023295] env[62569]: DEBUG oslo_vmware.rw_handles [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 950.023295] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52169769-4e82-75d3-1a54-e1c24183716e" [ 950.023295] env[62569]: _type = "HttpNfcLease" [ 950.023295] env[62569]: }. {{(pid=62569) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 950.023886] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fc02a7c-94dc-4bd6-b31a-0b5d0a970887 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.031324] env[62569]: DEBUG oslo_vmware.rw_handles [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528202bf-a4c3-5741-86a5-42e257dd6a11/disk-0.vmdk from lease info. {{(pid=62569) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 950.031587] env[62569]: DEBUG oslo_vmware.rw_handles [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Creating HTTP connection to write to file with size = 31665664 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528202bf-a4c3-5741-86a5-42e257dd6a11/disk-0.vmdk. {{(pid=62569) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 950.100318] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-5c42bc4f-cc85-4df1-821f-c1479b5ce73a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.219139] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Applying migration context for instance 09eeb937-aabc-418d-9fd3-3c7a72f4ef75 as it has an incoming, in-progress migration 5b4b811b-c7c1-4e18-b28c-14df9b5a62ed. Migration status is pre-migrating {{(pid=62569) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 950.219139] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Applying migration context for instance b47cd2d7-0cd2-41af-8ed1-a6dfca323516 as it has an incoming, in-progress migration 54fd4982-eb4e-4229-a3e0-21648ce18eb7. Migration status is migrating {{(pid=62569) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 950.220494] env[62569]: INFO nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Updating resource usage from migration 5b4b811b-c7c1-4e18-b28c-14df9b5a62ed [ 950.220826] env[62569]: INFO nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Updating resource usage from migration 54fd4982-eb4e-4229-a3e0-21648ce18eb7 [ 950.237555] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96a3fac0-38fa-4ea4-9615-186a59eedf19 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.244267] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance cb9d8991-b7ba-4bf7-9d22-1391f4cea768 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 950.244451] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance e4eadcdf-a04a-4255-ba1c-fe20156c655f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 950.244581] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 534c9ece-c8df-4528-83d8-c6c0ef8c4793 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 950.244734] env[62569]: WARNING nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 950.244871] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 188cac03-4034-4a02-973a-fb1906399fd1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 950.244999] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 98089a34-074b-4bdb-92ae-f9e23d2551f1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 950.245133] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance af26f440-3515-4be7-9a03-8a0404c152d9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 950.245261] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance b6e5eefc-8c06-445b-a3af-9404578b6179 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 950.245377] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance d1fdff70-0401-49ca-bbbb-ef5eda266c9c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 950.245497] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 950.245608] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Migration 54fd4982-eb4e-4229-a3e0-21648ce18eb7 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 950.245733] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance b47cd2d7-0cd2-41af-8ed1-a6dfca323516 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 950.245855] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Migration 5b4b811b-c7c1-4e18-b28c-14df9b5a62ed is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 950.245963] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 09eeb937-aabc-418d-9fd3-3c7a72f4ef75 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 950.263051] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Updating instance 'b47cd2d7-0cd2-41af-8ed1-a6dfca323516' progress to 0 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 950.269791] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance abf56021-ddf8-4677-9e90-46dc96e89133 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 950.270149] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=62569) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 950.270224] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3136MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=62569) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 950.359409] env[62569]: DEBUG nova.network.neutron [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Updated VIF entry in instance network info cache for port 6ca4b28a-1255-4f88-a7b6-54b0378768b1. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 950.359838] env[62569]: DEBUG nova.network.neutron [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Updating instance_info_cache with network_info: [{"id": "6ca4b28a-1255-4f88-a7b6-54b0378768b1", "address": "fa:16:3e:03:e7:ad", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ca4b28a-12", "ovs_interfaceid": "6ca4b28a-1255-4f88-a7b6-54b0378768b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.497107] env[62569]: DEBUG oslo_vmware.api [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250440, 'name': RemoveSnapshot_Task, 'duration_secs': 0.908655} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.497107] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Deleted Snapshot of the VM instance {{(pid=62569) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 950.592715] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ba36640-3550-40b8-98a0-c8f464d52c1d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.600950] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-483b3965-6f96-466a-92f6-423dd7bf2a07 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.633638] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3316de65-a53c-4880-a720-557538f271d1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.641883] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4134b5c0-a50c-45bf-8ee0-5267a939906c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.659466] env[62569]: DEBUG nova.compute.provider_tree [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.734902] env[62569]: DEBUG nova.network.neutron [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Updating instance_info_cache with network_info: [{"id": "c967324f-ab14-480b-a8b8-710db5827863", "address": "fa:16:3e:f9:c6:da", "network": {"id": "c61b7867-7897-43ae-8f5b-19834c4390ea", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-512066189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ba7e96d226942bd99ab29fe703c6fea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc967324f-ab", "ovs_interfaceid": "c967324f-ab14-480b-a8b8-710db5827863", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.779937] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 950.783009] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a61efbf9-6863-4bc8-91dd-63868b481fea {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.791656] env[62569]: DEBUG oslo_vmware.api [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 950.791656] env[62569]: value = "task-1250442" [ 950.791656] env[62569]: _type = "Task" [ 950.791656] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.802233] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] VM already powered off {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 950.802505] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Updating instance 'b47cd2d7-0cd2-41af-8ed1-a6dfca323516' progress to 17 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 950.864448] env[62569]: DEBUG oslo_concurrency.lockutils [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] Releasing lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.864797] env[62569]: DEBUG nova.compute.manager [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Received event network-changed-91986023-125e-442b-b105-102630ec6036 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 950.865032] env[62569]: DEBUG nova.compute.manager [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Refreshing instance network info cache due to event network-changed-91986023-125e-442b-b105-102630ec6036. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 950.865261] env[62569]: DEBUG oslo_concurrency.lockutils [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] Acquiring lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.865427] env[62569]: DEBUG oslo_concurrency.lockutils [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] Acquired lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.865655] env[62569]: DEBUG nova.network.neutron [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Refreshing network info cache for port 91986023-125e-442b-b105-102630ec6036 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 951.002103] env[62569]: WARNING nova.compute.manager [None req-5d9a7616-1b22-47e9-aecb-1b7daae275ce tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Image not found during snapshot: nova.exception.ImageNotFound: Image dcad6631-f683-4ed5-9957-aac1d800c83e could not be found. [ 951.162059] env[62569]: DEBUG nova.scheduler.client.report [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 951.240629] env[62569]: DEBUG oslo_concurrency.lockutils [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Releasing lock "refresh_cache-09eeb937-aabc-418d-9fd3-3c7a72f4ef75" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.310052] env[62569]: DEBUG nova.virt.hardware [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 951.310301] env[62569]: DEBUG nova.virt.hardware [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 951.310395] env[62569]: DEBUG nova.virt.hardware [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 951.310652] env[62569]: DEBUG nova.virt.hardware [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 951.310838] env[62569]: DEBUG nova.virt.hardware [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 951.310998] env[62569]: DEBUG nova.virt.hardware [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 951.311380] env[62569]: DEBUG nova.virt.hardware [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 951.311601] env[62569]: DEBUG nova.virt.hardware [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 951.311852] env[62569]: DEBUG nova.virt.hardware [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 951.312065] env[62569]: DEBUG nova.virt.hardware [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 951.312275] env[62569]: DEBUG nova.virt.hardware [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 951.320848] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45ed1bcf-c82f-4b6d-83b2-9929b0fd4591 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.340763] env[62569]: DEBUG oslo_vmware.api [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 951.340763] env[62569]: value = "task-1250443" [ 951.340763] env[62569]: _type = "Task" [ 951.340763] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.349150] env[62569]: DEBUG oslo_vmware.api [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250443, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.359768] env[62569]: DEBUG oslo_vmware.rw_handles [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Completed reading data from the image iterator. {{(pid=62569) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 951.360019] env[62569]: DEBUG oslo_vmware.rw_handles [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528202bf-a4c3-5741-86a5-42e257dd6a11/disk-0.vmdk. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 951.360991] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77bf6419-397c-4838-a41f-c270d9fc70fe {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.367296] env[62569]: DEBUG oslo_vmware.rw_handles [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528202bf-a4c3-5741-86a5-42e257dd6a11/disk-0.vmdk is in state: ready. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 951.367500] env[62569]: DEBUG oslo_vmware.rw_handles [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528202bf-a4c3-5741-86a5-42e257dd6a11/disk-0.vmdk. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 951.369692] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-a1a454a7-907d-47e6-a03e-5d2f3721a693 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.480126] env[62569]: DEBUG oslo_vmware.rw_handles [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fbe02f-5904-61cd-6beb-d306008f74fd/disk-0.vmdk. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 951.481451] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-601b5d51-ddd0-4807-bdb1-63d111dc6740 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.489521] env[62569]: DEBUG oslo_vmware.rw_handles [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fbe02f-5904-61cd-6beb-d306008f74fd/disk-0.vmdk is in state: ready. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 951.489765] env[62569]: ERROR oslo_vmware.rw_handles [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fbe02f-5904-61cd-6beb-d306008f74fd/disk-0.vmdk due to incomplete transfer. [ 951.490039] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f54d28ef-eb1c-4253-899f-0f4421903f79 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.506787] env[62569]: DEBUG oslo_vmware.rw_handles [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52fbe02f-5904-61cd-6beb-d306008f74fd/disk-0.vmdk. {{(pid=62569) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 951.507059] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Uploaded image deff6cce-4781-495b-a37b-642db84153cd to the Glance image server {{(pid=62569) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 951.509450] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Destroying the VM {{(pid=62569) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 951.512030] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4857ecd3-421c-4d85-a952-62c39f9f9233 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.517909] env[62569]: DEBUG oslo_vmware.api [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 951.517909] env[62569]: value = "task-1250444" [ 951.517909] env[62569]: _type = "Task" [ 951.517909] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.527280] env[62569]: DEBUG oslo_vmware.api [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250444, 'name': Destroy_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.648301] env[62569]: DEBUG nova.network.neutron [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Updated VIF entry in instance network info cache for port 91986023-125e-442b-b105-102630ec6036. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 951.648700] env[62569]: DEBUG nova.network.neutron [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Updating instance_info_cache with network_info: [{"id": "91986023-125e-442b-b105-102630ec6036", "address": "fa:16:3e:18:83:9b", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91986023-12", "ovs_interfaceid": "91986023-125e-442b-b105-102630ec6036", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.668053] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62569) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 951.668053] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.468s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.668270] env[62569]: DEBUG oslo_concurrency.lockutils [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.850s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.668325] env[62569]: DEBUG oslo_concurrency.lockutils [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.670398] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.275s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.672046] env[62569]: INFO nova.compute.claims [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 951.702117] env[62569]: INFO nova.scheduler.client.report [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Deleted allocations for instance bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a [ 951.852294] env[62569]: DEBUG oslo_vmware.api [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250443, 'name': ReconfigVM_Task, 'duration_secs': 0.186849} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.852685] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Updating instance 'b47cd2d7-0cd2-41af-8ed1-a6dfca323516' progress to 33 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 951.864160] env[62569]: DEBUG oslo_vmware.rw_handles [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528202bf-a4c3-5741-86a5-42e257dd6a11/disk-0.vmdk. {{(pid=62569) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 951.864393] env[62569]: INFO nova.virt.vmwareapi.images [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Downloaded image file data 87099cdd-38b2-4615-944d-a91a20479717 [ 951.865241] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85c5747f-0645-4ba3-be2d-c43a5bf71c79 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.880717] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-32e89b8a-4199-43ee-b479-0a6d62da48cb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.916646] env[62569]: INFO nova.virt.vmwareapi.images [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] The imported VM was unregistered [ 951.918644] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Caching image {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 951.918892] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Creating directory with path [datastore1] devstack-image-cache_base/87099cdd-38b2-4615-944d-a91a20479717 {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 951.919303] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e127e2f2-8962-464e-95bd-4b247ca3f385 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.929373] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Created directory with path [datastore1] devstack-image-cache_base/87099cdd-38b2-4615-944d-a91a20479717 {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 951.929561] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_6a50e58e-0561-481e-bf5f-89bb6d1df7d9/OSTACK_IMG_6a50e58e-0561-481e-bf5f-89bb6d1df7d9.vmdk to [datastore1] devstack-image-cache_base/87099cdd-38b2-4615-944d-a91a20479717/87099cdd-38b2-4615-944d-a91a20479717.vmdk. {{(pid=62569) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 951.929802] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-420d0d9d-2829-452a-ab8f-fe8ea5fa3c45 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.936089] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 951.936089] env[62569]: value = "task-1250446" [ 951.936089] env[62569]: _type = "Task" [ 951.936089] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.943332] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250446, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.029366] env[62569]: DEBUG oslo_vmware.api [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250444, 'name': Destroy_Task, 'duration_secs': 0.379576} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.029645] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Destroyed the VM [ 952.029892] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Deleting Snapshot of the VM instance {{(pid=62569) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 952.030162] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0fad8bfa-91b1-46bf-99e6-687690b3bc98 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.036587] env[62569]: DEBUG oslo_vmware.api [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 952.036587] env[62569]: value = "task-1250447" [ 952.036587] env[62569]: _type = "Task" [ 952.036587] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.044671] env[62569]: DEBUG oslo_vmware.api [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250447, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.151420] env[62569]: DEBUG oslo_concurrency.lockutils [req-d088c901-0d20-4b3d-b588-fad6d8715328 req-ed83cebf-bc79-4c11-bca8-2c1de9ff9238 service nova] Releasing lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.211599] env[62569]: DEBUG oslo_concurrency.lockutils [None req-100727b5-71e5-4799-8670-c2aa46a9b8aa tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.570s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.361602] env[62569]: DEBUG nova.virt.hardware [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 952.362079] env[62569]: DEBUG nova.virt.hardware [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 952.362416] env[62569]: DEBUG nova.virt.hardware [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 952.362664] env[62569]: DEBUG nova.virt.hardware [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 952.362822] env[62569]: DEBUG nova.virt.hardware [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 952.363049] env[62569]: DEBUG nova.virt.hardware [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 952.363225] env[62569]: DEBUG nova.virt.hardware [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 952.363418] env[62569]: DEBUG nova.virt.hardware [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 952.364035] env[62569]: DEBUG nova.virt.hardware [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 952.364035] env[62569]: DEBUG nova.virt.hardware [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 952.364189] env[62569]: DEBUG nova.virt.hardware [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 952.371079] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Reconfiguring VM instance instance-00000047 to detach disk 2000 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 952.371402] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c697be75-3619-4cad-9d0f-50e47fa485d5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.392494] env[62569]: DEBUG oslo_vmware.api [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 952.392494] env[62569]: value = "task-1250448" [ 952.392494] env[62569]: _type = "Task" [ 952.392494] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.402197] env[62569]: DEBUG oslo_vmware.api [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250448, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.445680] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250446, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.513112] env[62569]: DEBUG oslo_concurrency.lockutils [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "d1fdff70-0401-49ca-bbbb-ef5eda266c9c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.513444] env[62569]: DEBUG oslo_concurrency.lockutils [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "d1fdff70-0401-49ca-bbbb-ef5eda266c9c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.513679] env[62569]: DEBUG oslo_concurrency.lockutils [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "d1fdff70-0401-49ca-bbbb-ef5eda266c9c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.513891] env[62569]: DEBUG oslo_concurrency.lockutils [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "d1fdff70-0401-49ca-bbbb-ef5eda266c9c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.514490] env[62569]: DEBUG oslo_concurrency.lockutils [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "d1fdff70-0401-49ca-bbbb-ef5eda266c9c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.516595] env[62569]: INFO nova.compute.manager [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Terminating instance [ 952.546564] env[62569]: DEBUG oslo_vmware.api [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250447, 'name': RemoveSnapshot_Task, 'duration_secs': 0.481502} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.546909] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Deleted Snapshot of the VM instance {{(pid=62569) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 952.547239] env[62569]: DEBUG nova.compute.manager [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 952.548184] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aa3fe89-2e68-41d9-ac7f-52458e53edf0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.761693] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f4acc7b-a38b-459d-879a-33d6e43b45b4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.781875] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Updating instance '09eeb937-aabc-418d-9fd3-3c7a72f4ef75' progress to 0 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 952.904603] env[62569]: DEBUG oslo_vmware.api [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250448, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.931070] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb0c61a-83b8-4332-9e25-40097a33651e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.942475] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-303ca98d-e615-43a0-add9-61664f569fa0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.951187] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250446, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.980060] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c98f42f-eabf-4d46-9cfc-7cc3156e12bd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.988239] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c89d1304-28cf-431b-b44d-57205db4cca5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.798081] env[62569]: DEBUG nova.compute.manager [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 953.798483] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 953.802513] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 953.802939] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ec09ed5d-40bf-4c4b-b351-65bd874411c6 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "cb9d8991-b7ba-4bf7-9d22-1391f4cea768" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.803184] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ec09ed5d-40bf-4c4b-b351-65bd874411c6 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cb9d8991-b7ba-4bf7-9d22-1391f4cea768" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.808917] env[62569]: INFO nova.compute.manager [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Shelve offloading [ 953.811441] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6856f393-8ca5-45b5-9508-ee2b581b66dd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.815060] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a3d3b3c-729f-4c61-914e-92d7a46974d8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.822803] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "bcc7f73b-083b-477e-bceb-4f984a95d219" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.823184] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "bcc7f73b-083b-477e-bceb-4f984a95d219" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.835604] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "interface-188cac03-4034-4a02-973a-fb1906399fd1-0d27f2d0-6f98-4031-97c8-33df605f1bb0" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.835847] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "interface-188cac03-4034-4a02-973a-fb1906399fd1-0d27f2d0-6f98-4031-97c8-33df605f1bb0" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.836417] env[62569]: DEBUG nova.objects.instance [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lazy-loading 'flavor' on Instance uuid 188cac03-4034-4a02-973a-fb1906399fd1 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.838029] env[62569]: DEBUG nova.compute.provider_tree [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.846823] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 953.854331] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bbedd17c-5b71-4d88-a54e-621f877fab89 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.858385] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250446, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.858695] env[62569]: DEBUG oslo_vmware.api [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250448, 'name': ReconfigVM_Task, 'duration_secs': 1.122756} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.861089] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Reconfigured VM instance instance-00000047 to detach disk 2000 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 953.861595] env[62569]: DEBUG oslo_vmware.api [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 953.861595] env[62569]: value = "task-1250449" [ 953.861595] env[62569]: _type = "Task" [ 953.861595] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.862820] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-017a86f7-7b93-4e44-998b-37359393ece6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.872910] env[62569]: DEBUG oslo_vmware.api [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 953.872910] env[62569]: value = "task-1250450" [ 953.872910] env[62569]: _type = "Task" [ 953.872910] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.899824] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] b47cd2d7-0cd2-41af-8ed1-a6dfca323516/b47cd2d7-0cd2-41af-8ed1-a6dfca323516.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 953.900259] env[62569]: DEBUG oslo_vmware.api [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250449, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.900989] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae66ff0f-139f-49d6-a87b-e38d74dc3057 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.918403] env[62569]: DEBUG oslo_vmware.api [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250450, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.923708] env[62569]: DEBUG oslo_vmware.api [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 953.923708] env[62569]: value = "task-1250451" [ 953.923708] env[62569]: _type = "Task" [ 953.923708] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.933992] env[62569]: DEBUG oslo_vmware.api [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250451, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.314322] env[62569]: INFO nova.compute.manager [None req-ec09ed5d-40bf-4c4b-b351-65bd874411c6 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Detaching volume f29f7945-6df2-4fc4-bd59-5d5daee539e7 [ 954.316364] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250446, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.340852] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 954.344526] env[62569]: DEBUG nova.compute.manager [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 954.349185] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-afabfd13-2712-42ae-b8b5-e39908b560f2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.351746] env[62569]: DEBUG nova.scheduler.client.report [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 954.359375] env[62569]: INFO nova.virt.block_device [None req-ec09ed5d-40bf-4c4b-b351-65bd874411c6 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Attempting to driver detach volume f29f7945-6df2-4fc4-bd59-5d5daee539e7 from mountpoint /dev/sdb [ 954.359736] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec09ed5d-40bf-4c4b-b351-65bd874411c6 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Volume detach. Driver type: vmdk {{(pid=62569) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 954.360329] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec09ed5d-40bf-4c4b-b351-65bd874411c6 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269446', 'volume_id': 'f29f7945-6df2-4fc4-bd59-5d5daee539e7', 'name': 'volume-f29f7945-6df2-4fc4-bd59-5d5daee539e7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cb9d8991-b7ba-4bf7-9d22-1391f4cea768', 'attached_at': '', 'detached_at': '', 'volume_id': 'f29f7945-6df2-4fc4-bd59-5d5daee539e7', 'serial': 'f29f7945-6df2-4fc4-bd59-5d5daee539e7'} {{(pid=62569) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 954.361613] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8d5dece-a764-4dee-aac4-ff8cb075284c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.366181] env[62569]: DEBUG oslo_vmware.api [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 954.366181] env[62569]: value = "task-1250452" [ 954.366181] env[62569]: _type = "Task" [ 954.366181] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.393903] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d0a73f7-c3ae-41bb-8546-362f083917dc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.403198] env[62569]: DEBUG oslo_vmware.api [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250449, 'name': PowerOffVM_Task, 'duration_secs': 0.34659} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.404520] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 954.404724] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Updating instance '09eeb937-aabc-418d-9fd3-3c7a72f4ef75' progress to 17 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 954.408797] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] VM already powered off {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 954.409134] env[62569]: DEBUG nova.compute.manager [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 954.417294] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a94553b0-f82c-48be-a17b-896cae76af3d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.421937] env[62569]: DEBUG oslo_vmware.api [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250450, 'name': PowerOffVM_Task, 'duration_secs': 0.340592} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.451181] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9678ee58-2c48-484c-91d4-24dcf9dfed8f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.451181] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 954.451181] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 954.451181] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cf0500bc-5c45-4269-ba2e-651447d9fd68 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.451181] env[62569]: DEBUG oslo_concurrency.lockutils [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "refresh_cache-af26f440-3515-4be7-9a03-8a0404c152d9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.451181] env[62569]: DEBUG oslo_concurrency.lockutils [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired lock "refresh_cache-af26f440-3515-4be7-9a03-8a0404c152d9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.451181] env[62569]: DEBUG nova.network.neutron [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 954.473118] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e8db47a-75bc-483d-851e-0200bb864765 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.480409] env[62569]: DEBUG oslo_vmware.api [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250451, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.494816] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec09ed5d-40bf-4c4b-b351-65bd874411c6 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] The volume has not been displaced from its original location: [datastore2] volume-f29f7945-6df2-4fc4-bd59-5d5daee539e7/volume-f29f7945-6df2-4fc4-bd59-5d5daee539e7.vmdk. No consolidation needed. {{(pid=62569) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 954.500342] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec09ed5d-40bf-4c4b-b351-65bd874411c6 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Reconfiguring VM instance instance-00000048 to detach disk 2001 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 954.500551] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b746885d-55b7-46ab-880f-74a6dcb243ed {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.522112] env[62569]: DEBUG oslo_vmware.api [None req-ec09ed5d-40bf-4c4b-b351-65bd874411c6 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 954.522112] env[62569]: value = "task-1250454" [ 954.522112] env[62569]: _type = "Task" [ 954.522112] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.530846] env[62569]: DEBUG oslo_vmware.api [None req-ec09ed5d-40bf-4c4b-b351-65bd874411c6 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250454, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.543967] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 954.544206] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 954.544391] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Deleting the datastore file [datastore2] d1fdff70-0401-49ca-bbbb-ef5eda266c9c {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 954.544672] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e3808fa3-d94b-4ab5-b897-556ba0e4f5f4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.550723] env[62569]: DEBUG oslo_vmware.api [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for the task: (returnval){ [ 954.550723] env[62569]: value = "task-1250455" [ 954.550723] env[62569]: _type = "Task" [ 954.550723] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.559023] env[62569]: DEBUG oslo_vmware.api [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250455, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.731107] env[62569]: DEBUG nova.network.neutron [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Updating instance_info_cache with network_info: [{"id": "9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c", "address": "fa:16:3e:75:9a:a6", "network": {"id": "30d3bf4a-8f55-492f-a049-181e29731818", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1244949475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bdba6022e3d4697a336ca28ca4eccec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c1cf99c-fb", "ovs_interfaceid": "9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.764651] env[62569]: DEBUG nova.objects.instance [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lazy-loading 'pci_requests' on Instance uuid 188cac03-4034-4a02-973a-fb1906399fd1 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.816394] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250446, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.587653} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.816719] env[62569]: INFO nova.virt.vmwareapi.ds_util [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_6a50e58e-0561-481e-bf5f-89bb6d1df7d9/OSTACK_IMG_6a50e58e-0561-481e-bf5f-89bb6d1df7d9.vmdk to [datastore1] devstack-image-cache_base/87099cdd-38b2-4615-944d-a91a20479717/87099cdd-38b2-4615-944d-a91a20479717.vmdk. [ 954.816870] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Cleaning up location [datastore1] OSTACK_IMG_6a50e58e-0561-481e-bf5f-89bb6d1df7d9 {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 954.817050] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_6a50e58e-0561-481e-bf5f-89bb6d1df7d9 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 954.817325] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-31a006bf-5695-4251-aac6-b2885cf082f2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.824180] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 954.824180] env[62569]: value = "task-1250456" [ 954.824180] env[62569]: _type = "Task" [ 954.824180] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.832831] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250456, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.858843] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.188s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.860039] env[62569]: DEBUG nova.compute.manager [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 954.889026] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.889454] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.892352] env[62569]: INFO nova.compute.claims [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 954.912290] env[62569]: DEBUG nova.virt.hardware [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 954.912564] env[62569]: DEBUG nova.virt.hardware [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 954.912770] env[62569]: DEBUG nova.virt.hardware [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 954.913016] env[62569]: DEBUG nova.virt.hardware [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 954.913200] env[62569]: DEBUG nova.virt.hardware [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 954.913375] env[62569]: DEBUG nova.virt.hardware [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 954.913582] env[62569]: DEBUG nova.virt.hardware [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 954.913749] env[62569]: DEBUG nova.virt.hardware [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 954.913924] env[62569]: DEBUG nova.virt.hardware [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 954.914108] env[62569]: DEBUG nova.virt.hardware [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 954.914302] env[62569]: DEBUG nova.virt.hardware [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 954.919723] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b7a8a1c-bd94-4941-9210-cc34a60d801e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.940749] env[62569]: DEBUG oslo_vmware.api [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250451, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.942246] env[62569]: DEBUG oslo_vmware.api [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 954.942246] env[62569]: value = "task-1250457" [ 954.942246] env[62569]: _type = "Task" [ 954.942246] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.950674] env[62569]: DEBUG oslo_vmware.api [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250457, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.031968] env[62569]: DEBUG oslo_vmware.api [None req-ec09ed5d-40bf-4c4b-b351-65bd874411c6 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250454, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.059306] env[62569]: DEBUG oslo_vmware.api [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250455, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.234457] env[62569]: DEBUG oslo_concurrency.lockutils [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Releasing lock "refresh_cache-af26f440-3515-4be7-9a03-8a0404c152d9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.267206] env[62569]: DEBUG nova.objects.base [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Object Instance<188cac03-4034-4a02-973a-fb1906399fd1> lazy-loaded attributes: flavor,pci_requests {{(pid=62569) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 955.267458] env[62569]: DEBUG nova.network.neutron [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 955.335107] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250456, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.044294} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.336433] env[62569]: DEBUG nova.policy [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ab6d20862c54432cbafdda33ccc974ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '530ca7e7924743ab91a362a064a3111b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 955.338264] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 955.338522] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Releasing lock "[datastore1] devstack-image-cache_base/87099cdd-38b2-4615-944d-a91a20479717/87099cdd-38b2-4615-944d-a91a20479717.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.338826] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/87099cdd-38b2-4615-944d-a91a20479717/87099cdd-38b2-4615-944d-a91a20479717.vmdk to [datastore1] b6e5eefc-8c06-445b-a3af-9404578b6179/b6e5eefc-8c06-445b-a3af-9404578b6179.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 955.339143] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8cc9dc61-ca37-42a8-a8d1-d08387637173 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.346305] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 955.346305] env[62569]: value = "task-1250458" [ 955.346305] env[62569]: _type = "Task" [ 955.346305] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.353763] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250458, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.367271] env[62569]: DEBUG nova.compute.utils [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 955.368912] env[62569]: DEBUG nova.compute.manager [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 955.369384] env[62569]: DEBUG nova.network.neutron [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 955.440438] env[62569]: DEBUG nova.policy [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6dd052494f244fd78050c5d26f497720', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2c6a4814c234d53bbd4844cd482fbae', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 955.450615] env[62569]: DEBUG oslo_vmware.api [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250451, 'name': ReconfigVM_Task, 'duration_secs': 1.116967} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.451264] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Reconfigured VM instance instance-00000047 to attach disk [datastore1] b47cd2d7-0cd2-41af-8ed1-a6dfca323516/b47cd2d7-0cd2-41af-8ed1-a6dfca323516.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 955.451589] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Updating instance 'b47cd2d7-0cd2-41af-8ed1-a6dfca323516' progress to 50 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 955.458580] env[62569]: DEBUG oslo_vmware.api [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250457, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.529941] env[62569]: DEBUG nova.compute.manager [req-99b83702-d10b-4f6e-90e1-a10d2c59352f req-061f2374-a5b7-428b-86eb-ff74a500ec2b service nova] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Received event network-vif-unplugged-9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 955.530278] env[62569]: DEBUG oslo_concurrency.lockutils [req-99b83702-d10b-4f6e-90e1-a10d2c59352f req-061f2374-a5b7-428b-86eb-ff74a500ec2b service nova] Acquiring lock "af26f440-3515-4be7-9a03-8a0404c152d9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.530563] env[62569]: DEBUG oslo_concurrency.lockutils [req-99b83702-d10b-4f6e-90e1-a10d2c59352f req-061f2374-a5b7-428b-86eb-ff74a500ec2b service nova] Lock "af26f440-3515-4be7-9a03-8a0404c152d9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.530744] env[62569]: DEBUG oslo_concurrency.lockutils [req-99b83702-d10b-4f6e-90e1-a10d2c59352f req-061f2374-a5b7-428b-86eb-ff74a500ec2b service nova] Lock "af26f440-3515-4be7-9a03-8a0404c152d9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.530946] env[62569]: DEBUG nova.compute.manager [req-99b83702-d10b-4f6e-90e1-a10d2c59352f req-061f2374-a5b7-428b-86eb-ff74a500ec2b service nova] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] No waiting events found dispatching network-vif-unplugged-9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 955.531131] env[62569]: WARNING nova.compute.manager [req-99b83702-d10b-4f6e-90e1-a10d2c59352f req-061f2374-a5b7-428b-86eb-ff74a500ec2b service nova] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Received unexpected event network-vif-unplugged-9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c for instance with vm_state shelved and task_state shelving_offloading. [ 955.538743] env[62569]: DEBUG oslo_vmware.api [None req-ec09ed5d-40bf-4c4b-b351-65bd874411c6 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250454, 'name': ReconfigVM_Task, 'duration_secs': 0.561963} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.539016] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec09ed5d-40bf-4c4b-b351-65bd874411c6 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Reconfigured VM instance instance-00000048 to detach disk 2001 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 955.545471] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23f9d619-7759-4ca6-8bd9-aa9c3cc24934 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.556835] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 955.558320] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd809d0e-8026-466b-9d21-e0d5f87e5a3d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.569399] env[62569]: DEBUG oslo_vmware.api [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Task: {'id': task-1250455, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.527084} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.573093] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 955.573305] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 955.573488] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 955.573661] env[62569]: INFO nova.compute.manager [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Took 1.78 seconds to destroy the instance on the hypervisor. [ 955.573898] env[62569]: DEBUG oslo.service.loopingcall [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 955.574304] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 955.574608] env[62569]: DEBUG oslo_vmware.api [None req-ec09ed5d-40bf-4c4b-b351-65bd874411c6 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 955.574608] env[62569]: value = "task-1250459" [ 955.574608] env[62569]: _type = "Task" [ 955.574608] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.575072] env[62569]: DEBUG nova.compute.manager [-] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 955.575171] env[62569]: DEBUG nova.network.neutron [-] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 955.576977] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2f211c31-bff7-4260-a23f-1951de58beeb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.587607] env[62569]: DEBUG oslo_vmware.api [None req-ec09ed5d-40bf-4c4b-b351-65bd874411c6 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250459, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.643626] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 955.643892] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 955.644170] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Deleting the datastore file [datastore2] af26f440-3515-4be7-9a03-8a0404c152d9 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 955.644519] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8a186990-c9b6-4d90-91e4-6008be27206f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.652990] env[62569]: DEBUG oslo_vmware.api [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 955.652990] env[62569]: value = "task-1250461" [ 955.652990] env[62569]: _type = "Task" [ 955.652990] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.665809] env[62569]: DEBUG oslo_vmware.api [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250461, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.832175] env[62569]: DEBUG nova.network.neutron [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Successfully created port: 0349ff40-8656-4ada-9f99-9196782155c6 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 955.859652] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250458, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.873569] env[62569]: DEBUG nova.compute.manager [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 955.958012] env[62569]: DEBUG oslo_vmware.api [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250457, 'name': ReconfigVM_Task, 'duration_secs': 0.536174} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.959019] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Updating instance '09eeb937-aabc-418d-9fd3-3c7a72f4ef75' progress to 33 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 955.971356] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76b078a8-7b4f-4755-9500-ce06c08e5280 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.996226] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ff266e-1682-4540-a540-40d139ec47b3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.017964] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Updating instance 'b47cd2d7-0cd2-41af-8ed1-a6dfca323516' progress to 67 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 956.088916] env[62569]: DEBUG oslo_vmware.api [None req-ec09ed5d-40bf-4c4b-b351-65bd874411c6 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250459, 'name': ReconfigVM_Task, 'duration_secs': 0.16604} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.092044] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec09ed5d-40bf-4c4b-b351-65bd874411c6 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269446', 'volume_id': 'f29f7945-6df2-4fc4-bd59-5d5daee539e7', 'name': 'volume-f29f7945-6df2-4fc4-bd59-5d5daee539e7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cb9d8991-b7ba-4bf7-9d22-1391f4cea768', 'attached_at': '', 'detached_at': '', 'volume_id': 'f29f7945-6df2-4fc4-bd59-5d5daee539e7', 'serial': 'f29f7945-6df2-4fc4-bd59-5d5daee539e7'} {{(pid=62569) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 956.165462] env[62569]: DEBUG oslo_vmware.api [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250461, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.307142} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.166860] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.167093] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 956.167284] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 956.169792] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67991319-b56e-4b57-8ed1-50797e03d38d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.177394] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c53cb829-ab0c-429a-9627-5de9863c2a60 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.209135] env[62569]: INFO nova.scheduler.client.report [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Deleted allocations for instance af26f440-3515-4be7-9a03-8a0404c152d9 [ 956.213339] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70a0a1ab-7322-4c6f-b140-50d67fa00de1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.221813] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-566c329e-2939-47eb-a472-d7f46f63edc9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.235377] env[62569]: DEBUG nova.compute.provider_tree [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 956.359017] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250458, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.397485] env[62569]: DEBUG nova.network.neutron [-] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.469510] env[62569]: DEBUG nova.virt.hardware [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 956.469869] env[62569]: DEBUG nova.virt.hardware [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 956.470164] env[62569]: DEBUG nova.virt.hardware [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 956.470304] env[62569]: DEBUG nova.virt.hardware [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 956.470461] env[62569]: DEBUG nova.virt.hardware [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 956.470641] env[62569]: DEBUG nova.virt.hardware [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 956.470888] env[62569]: DEBUG nova.virt.hardware [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 956.471071] env[62569]: DEBUG nova.virt.hardware [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 956.471252] env[62569]: DEBUG nova.virt.hardware [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 956.471515] env[62569]: DEBUG nova.virt.hardware [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 956.471708] env[62569]: DEBUG nova.virt.hardware [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 956.477315] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Reconfiguring VM instance instance-0000005b to detach disk 2000 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 956.478030] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8a6002ec-aa63-43e9-aae6-09baac8e572b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.497290] env[62569]: DEBUG oslo_vmware.api [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 956.497290] env[62569]: value = "task-1250462" [ 956.497290] env[62569]: _type = "Task" [ 956.497290] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.505608] env[62569]: DEBUG oslo_vmware.api [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250462, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.586500] env[62569]: DEBUG nova.network.neutron [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Port 90b3aaef-2e72-486f-be8d-a4202a6cb0c6 binding to destination host cpu-1 is already ACTIVE {{(pid=62569) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 956.638017] env[62569]: DEBUG nova.objects.instance [None req-ec09ed5d-40bf-4c4b-b351-65bd874411c6 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lazy-loading 'flavor' on Instance uuid cb9d8991-b7ba-4bf7-9d22-1391f4cea768 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 956.718668] env[62569]: DEBUG oslo_concurrency.lockutils [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.757212] env[62569]: ERROR nova.scheduler.client.report [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [req-2eb091aa-a72f-4265-b77f-9fbd664c95f6] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID fa06556a-5785-4014-b8bd-bc240a0cf716. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2eb091aa-a72f-4265-b77f-9fbd664c95f6"}]} [ 956.774503] env[62569]: DEBUG nova.scheduler.client.report [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Refreshing inventories for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 956.789743] env[62569]: DEBUG nova.scheduler.client.report [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Updating ProviderTree inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 956.789999] env[62569]: DEBUG nova.compute.provider_tree [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 956.804124] env[62569]: DEBUG nova.scheduler.client.report [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Refreshing aggregate associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, aggregates: None {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 956.829873] env[62569]: DEBUG nova.scheduler.client.report [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Refreshing trait associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 956.858779] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250458, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.883467] env[62569]: DEBUG nova.compute.manager [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 956.900288] env[62569]: INFO nova.compute.manager [-] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Took 1.32 seconds to deallocate network for instance. [ 956.917912] env[62569]: DEBUG nova.compute.manager [req-a2cd207f-a76d-483b-8348-921df53d92d5 req-aee405a9-892b-480a-8b65-b0635cd9f58a service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Received event network-vif-plugged-0d27f2d0-6f98-4031-97c8-33df605f1bb0 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 956.918165] env[62569]: DEBUG oslo_concurrency.lockutils [req-a2cd207f-a76d-483b-8348-921df53d92d5 req-aee405a9-892b-480a-8b65-b0635cd9f58a service nova] Acquiring lock "188cac03-4034-4a02-973a-fb1906399fd1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.918381] env[62569]: DEBUG oslo_concurrency.lockutils [req-a2cd207f-a76d-483b-8348-921df53d92d5 req-aee405a9-892b-480a-8b65-b0635cd9f58a service nova] Lock "188cac03-4034-4a02-973a-fb1906399fd1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.918555] env[62569]: DEBUG oslo_concurrency.lockutils [req-a2cd207f-a76d-483b-8348-921df53d92d5 req-aee405a9-892b-480a-8b65-b0635cd9f58a service nova] Lock "188cac03-4034-4a02-973a-fb1906399fd1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.918726] env[62569]: DEBUG nova.compute.manager [req-a2cd207f-a76d-483b-8348-921df53d92d5 req-aee405a9-892b-480a-8b65-b0635cd9f58a service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] No waiting events found dispatching network-vif-plugged-0d27f2d0-6f98-4031-97c8-33df605f1bb0 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 956.918893] env[62569]: WARNING nova.compute.manager [req-a2cd207f-a76d-483b-8348-921df53d92d5 req-aee405a9-892b-480a-8b65-b0635cd9f58a service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Received unexpected event network-vif-plugged-0d27f2d0-6f98-4031-97c8-33df605f1bb0 for instance with vm_state active and task_state None. [ 956.923830] env[62569]: DEBUG nova.virt.hardware [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 956.924103] env[62569]: DEBUG nova.virt.hardware [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 956.924305] env[62569]: DEBUG nova.virt.hardware [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 956.924468] env[62569]: DEBUG nova.virt.hardware [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 956.924630] env[62569]: DEBUG nova.virt.hardware [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 956.924778] env[62569]: DEBUG nova.virt.hardware [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 956.924989] env[62569]: DEBUG nova.virt.hardware [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 956.925171] env[62569]: DEBUG nova.virt.hardware [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 956.925342] env[62569]: DEBUG nova.virt.hardware [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 956.925510] env[62569]: DEBUG nova.virt.hardware [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 956.925686] env[62569]: DEBUG nova.virt.hardware [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 956.926939] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92853206-ca72-4010-a68d-d28c4b68641a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.940049] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144d57bf-764e-4758-bceb-660195e5c1b1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.007710] env[62569]: DEBUG oslo_vmware.api [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250462, 'name': ReconfigVM_Task, 'duration_secs': 0.205233} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.010784] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Reconfigured VM instance instance-0000005b to detach disk 2000 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 957.011966] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3f262cb-397f-4206-bcd1-37de5e530379 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.038393] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] 09eeb937-aabc-418d-9fd3-3c7a72f4ef75/09eeb937-aabc-418d-9fd3-3c7a72f4ef75.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 957.041700] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82b09330-b27d-46ff-bc1f-c83f28adc9ec {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.057090] env[62569]: DEBUG nova.network.neutron [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Successfully updated port: 0d27f2d0-6f98-4031-97c8-33df605f1bb0 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 957.064374] env[62569]: DEBUG oslo_vmware.api [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 957.064374] env[62569]: value = "task-1250463" [ 957.064374] env[62569]: _type = "Task" [ 957.064374] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.073819] env[62569]: DEBUG oslo_vmware.api [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250463, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.106133] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25ae671e-8a62-4fac-b035-18b50fcaeb96 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.118051] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e642fa57-3ed2-48a3-a33f-e39fbb9e7510 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.153334] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-999dda09-7a70-425a-b67b-7e75b00d235a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.164432] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63a73556-2c93-47aa-a96e-5ea042ee939b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.181440] env[62569]: DEBUG nova.compute.provider_tree [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 957.360306] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250458, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.414917] env[62569]: DEBUG oslo_concurrency.lockutils [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.420880] env[62569]: DEBUG nova.network.neutron [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Successfully updated port: 0349ff40-8656-4ada-9f99-9196782155c6 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 957.560661] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.561009] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.561075] env[62569]: DEBUG nova.network.neutron [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 957.574909] env[62569]: DEBUG oslo_vmware.api [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250463, 'name': ReconfigVM_Task, 'duration_secs': 0.429156} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.575234] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Reconfigured VM instance instance-0000005b to attach disk [datastore2] 09eeb937-aabc-418d-9fd3-3c7a72f4ef75/09eeb937-aabc-418d-9fd3-3c7a72f4ef75.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 957.575518] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Updating instance '09eeb937-aabc-418d-9fd3-3c7a72f4ef75' progress to 50 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 957.610457] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "b47cd2d7-0cd2-41af-8ed1-a6dfca323516-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.610672] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "b47cd2d7-0cd2-41af-8ed1-a6dfca323516-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.610844] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "b47cd2d7-0cd2-41af-8ed1-a6dfca323516-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.658668] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ec09ed5d-40bf-4c4b-b351-65bd874411c6 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cb9d8991-b7ba-4bf7-9d22-1391f4cea768" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.855s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.703178] env[62569]: DEBUG nova.compute.manager [req-b53aa264-0263-4f64-a252-88eecceb3de4 req-72272ec6-12ba-4f18-a3d7-afcb8f2f4b9e service nova] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Received event network-changed-9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 957.703399] env[62569]: DEBUG nova.compute.manager [req-b53aa264-0263-4f64-a252-88eecceb3de4 req-72272ec6-12ba-4f18-a3d7-afcb8f2f4b9e service nova] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Refreshing instance network info cache due to event network-changed-9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 957.703628] env[62569]: DEBUG oslo_concurrency.lockutils [req-b53aa264-0263-4f64-a252-88eecceb3de4 req-72272ec6-12ba-4f18-a3d7-afcb8f2f4b9e service nova] Acquiring lock "refresh_cache-af26f440-3515-4be7-9a03-8a0404c152d9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.703780] env[62569]: DEBUG oslo_concurrency.lockutils [req-b53aa264-0263-4f64-a252-88eecceb3de4 req-72272ec6-12ba-4f18-a3d7-afcb8f2f4b9e service nova] Acquired lock "refresh_cache-af26f440-3515-4be7-9a03-8a0404c152d9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.703949] env[62569]: DEBUG nova.network.neutron [req-b53aa264-0263-4f64-a252-88eecceb3de4 req-72272ec6-12ba-4f18-a3d7-afcb8f2f4b9e service nova] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Refreshing network info cache for port 9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 957.714330] env[62569]: DEBUG nova.scheduler.client.report [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Updated inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with generation 125 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 957.715030] env[62569]: DEBUG nova.compute.provider_tree [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Updating resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 generation from 125 to 126 during operation: update_inventory {{(pid=62569) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 957.715030] env[62569]: DEBUG nova.compute.provider_tree [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 957.859956] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250458, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.436723} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.860588] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/87099cdd-38b2-4615-944d-a91a20479717/87099cdd-38b2-4615-944d-a91a20479717.vmdk to [datastore1] b6e5eefc-8c06-445b-a3af-9404578b6179/b6e5eefc-8c06-445b-a3af-9404578b6179.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 957.861399] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-550f71a7-94e3-47d2-952d-48ea573b4b6f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.883219] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] b6e5eefc-8c06-445b-a3af-9404578b6179/b6e5eefc-8c06-445b-a3af-9404578b6179.vmdk or device None with type streamOptimized {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 957.883515] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c77a5f5b-0318-4812-a096-d736db05acf8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.903214] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 957.903214] env[62569]: value = "task-1250464" [ 957.903214] env[62569]: _type = "Task" [ 957.903214] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.911547] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250464, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.923827] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "refresh_cache-abf56021-ddf8-4677-9e90-46dc96e89133" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.923827] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired lock "refresh_cache-abf56021-ddf8-4677-9e90-46dc96e89133" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.923827] env[62569]: DEBUG nova.network.neutron [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 958.082131] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c69007f7-d0e8-484f-b4c5-e770f3885d7f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.100207] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fe3b5602-a4b4-4f7a-b454-ec0b615fe240 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "af26f440-3515-4be7-9a03-8a0404c152d9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.101223] env[62569]: WARNING nova.network.neutron [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] 29fcbfe8-35ce-4e8d-874b-2b14ba720adf already exists in list: networks containing: ['29fcbfe8-35ce-4e8d-874b-2b14ba720adf']. ignoring it [ 958.103644] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f18c1750-d0fc-4af5-b6c8-841859436da8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.125012] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Updating instance '09eeb937-aabc-418d-9fd3-3c7a72f4ef75' progress to 67 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 958.173059] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "cb9d8991-b7ba-4bf7-9d22-1391f4cea768" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.173345] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cb9d8991-b7ba-4bf7-9d22-1391f4cea768" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.173642] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "cb9d8991-b7ba-4bf7-9d22-1391f4cea768-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.173996] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cb9d8991-b7ba-4bf7-9d22-1391f4cea768-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.174317] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cb9d8991-b7ba-4bf7-9d22-1391f4cea768-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.177127] env[62569]: INFO nova.compute.manager [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Terminating instance [ 958.219900] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.330s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.220421] env[62569]: DEBUG nova.compute.manager [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 958.222912] env[62569]: DEBUG oslo_concurrency.lockutils [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.505s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.223153] env[62569]: DEBUG nova.objects.instance [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lazy-loading 'resources' on Instance uuid af26f440-3515-4be7-9a03-8a0404c152d9 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.414611] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250464, 'name': ReconfigVM_Task, 'duration_secs': 0.288259} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.414940] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Reconfigured VM instance instance-0000003d to attach disk [datastore1] b6e5eefc-8c06-445b-a3af-9404578b6179/b6e5eefc-8c06-445b-a3af-9404578b6179.vmdk or device None with type streamOptimized {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 958.415614] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-37dc2d8f-3c1a-460d-beb3-c9d619843f30 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.421574] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 958.421574] env[62569]: value = "task-1250465" [ 958.421574] env[62569]: _type = "Task" [ 958.421574] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.430603] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250465, 'name': Rename_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.459808] env[62569]: DEBUG nova.network.neutron [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 958.520857] env[62569]: DEBUG nova.network.neutron [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Updating instance_info_cache with network_info: [{"id": "91986023-125e-442b-b105-102630ec6036", "address": "fa:16:3e:18:83:9b", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91986023-12", "ovs_interfaceid": "91986023-125e-442b-b105-102630ec6036", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0d27f2d0-6f98-4031-97c8-33df605f1bb0", "address": "fa:16:3e:e3:fb:62", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d27f2d0-6f", "ovs_interfaceid": "0d27f2d0-6f98-4031-97c8-33df605f1bb0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.523240] env[62569]: DEBUG nova.network.neutron [req-b53aa264-0263-4f64-a252-88eecceb3de4 req-72272ec6-12ba-4f18-a3d7-afcb8f2f4b9e service nova] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Updated VIF entry in instance network info cache for port 9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 958.523547] env[62569]: DEBUG nova.network.neutron [req-b53aa264-0263-4f64-a252-88eecceb3de4 req-72272ec6-12ba-4f18-a3d7-afcb8f2f4b9e service nova] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Updating instance_info_cache with network_info: [{"id": "9c1cf99c-fbc3-4dc8-8cd2-72cc80538b1c", "address": "fa:16:3e:75:9a:a6", "network": {"id": "30d3bf4a-8f55-492f-a049-181e29731818", "bridge": null, "label": "tempest-DeleteServersTestJSON-1244949475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bdba6022e3d4697a336ca28ca4eccec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap9c1cf99c-fb", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.591031] env[62569]: DEBUG nova.network.neutron [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Updating instance_info_cache with network_info: [{"id": "0349ff40-8656-4ada-9f99-9196782155c6", "address": "fa:16:3e:25:ee:d5", "network": {"id": "884d3248-07f1-4cf7-8ceb-b94b34622384", "bridge": "br-int", "label": "tempest-ServersTestJSON-1882023778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2c6a4814c234d53bbd4844cd482fbae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0349ff40-86", "ovs_interfaceid": "0349ff40-8656-4ada-9f99-9196782155c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.658930] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.659240] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.659428] env[62569]: DEBUG nova.network.neutron [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 958.665167] env[62569]: DEBUG nova.network.neutron [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Port c967324f-ab14-480b-a8b8-710db5827863 binding to destination host cpu-1 is already ACTIVE {{(pid=62569) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 958.680407] env[62569]: DEBUG nova.compute.manager [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 958.680617] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 958.681519] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1adb561-532d-4802-af89-aa15869070dd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.689810] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 958.690054] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8c9fecca-d674-4ad4-8c46-258f7f9dcff9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.696043] env[62569]: DEBUG oslo_vmware.api [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 958.696043] env[62569]: value = "task-1250466" [ 958.696043] env[62569]: _type = "Task" [ 958.696043] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.705424] env[62569]: DEBUG oslo_vmware.api [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250466, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.726392] env[62569]: DEBUG nova.compute.utils [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 958.728291] env[62569]: DEBUG nova.objects.instance [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lazy-loading 'numa_topology' on Instance uuid af26f440-3515-4be7-9a03-8a0404c152d9 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.729552] env[62569]: DEBUG nova.compute.manager [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 958.729771] env[62569]: DEBUG nova.network.neutron [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 958.767297] env[62569]: DEBUG nova.policy [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ccb9b3f4e38e4e81819d22e781661015', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7f988acdb9cd4ca28ca8916c2ede2db5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 958.931986] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250465, 'name': Rename_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.002200] env[62569]: DEBUG nova.network.neutron [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Successfully created port: a274afa6-f296-447a-902c-0747076a59c9 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 959.023689] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.024541] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.024700] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.025258] env[62569]: DEBUG oslo_concurrency.lockutils [req-b53aa264-0263-4f64-a252-88eecceb3de4 req-72272ec6-12ba-4f18-a3d7-afcb8f2f4b9e service nova] Releasing lock "refresh_cache-af26f440-3515-4be7-9a03-8a0404c152d9" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.026096] env[62569]: DEBUG nova.compute.manager [req-b53aa264-0263-4f64-a252-88eecceb3de4 req-72272ec6-12ba-4f18-a3d7-afcb8f2f4b9e service nova] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Received event network-vif-deleted-e1885f70-c997-4345-8ff7-0927c947b6cf {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 959.027221] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a44a7b67-d32a-4dd6-8275-42a62a78bbc0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.045456] env[62569]: DEBUG nova.virt.hardware [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 959.045840] env[62569]: DEBUG nova.virt.hardware [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 959.045840] env[62569]: DEBUG nova.virt.hardware [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 959.046067] env[62569]: DEBUG nova.virt.hardware [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 959.046204] env[62569]: DEBUG nova.virt.hardware [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 959.046381] env[62569]: DEBUG nova.virt.hardware [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 959.046595] env[62569]: DEBUG nova.virt.hardware [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 959.046775] env[62569]: DEBUG nova.virt.hardware [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 959.047111] env[62569]: DEBUG nova.virt.hardware [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 959.047236] env[62569]: DEBUG nova.virt.hardware [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 959.047430] env[62569]: DEBUG nova.virt.hardware [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 959.053811] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Reconfiguring VM to attach interface {{(pid=62569) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 959.055231] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d31fb7a1-2d87-4df6-9fcf-3b41b2a8f17d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.068673] env[62569]: DEBUG nova.compute.manager [req-132d2419-fd3e-428b-a7bd-57db6b7f52e8 req-92d94843-0ed0-4f2a-abc5-2827d9880723 service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Received event network-changed-0d27f2d0-6f98-4031-97c8-33df605f1bb0 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 959.068870] env[62569]: DEBUG nova.compute.manager [req-132d2419-fd3e-428b-a7bd-57db6b7f52e8 req-92d94843-0ed0-4f2a-abc5-2827d9880723 service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Refreshing instance network info cache due to event network-changed-0d27f2d0-6f98-4031-97c8-33df605f1bb0. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 959.069104] env[62569]: DEBUG oslo_concurrency.lockutils [req-132d2419-fd3e-428b-a7bd-57db6b7f52e8 req-92d94843-0ed0-4f2a-abc5-2827d9880723 service nova] Acquiring lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.069297] env[62569]: DEBUG oslo_concurrency.lockutils [req-132d2419-fd3e-428b-a7bd-57db6b7f52e8 req-92d94843-0ed0-4f2a-abc5-2827d9880723 service nova] Acquired lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.069481] env[62569]: DEBUG nova.network.neutron [req-132d2419-fd3e-428b-a7bd-57db6b7f52e8 req-92d94843-0ed0-4f2a-abc5-2827d9880723 service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Refreshing network info cache for port 0d27f2d0-6f98-4031-97c8-33df605f1bb0 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 959.079630] env[62569]: DEBUG oslo_vmware.api [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 959.079630] env[62569]: value = "task-1250467" [ 959.079630] env[62569]: _type = "Task" [ 959.079630] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.088449] env[62569]: DEBUG oslo_vmware.api [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250467, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.094128] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Releasing lock "refresh_cache-abf56021-ddf8-4677-9e90-46dc96e89133" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.094456] env[62569]: DEBUG nova.compute.manager [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Instance network_info: |[{"id": "0349ff40-8656-4ada-9f99-9196782155c6", "address": "fa:16:3e:25:ee:d5", "network": {"id": "884d3248-07f1-4cf7-8ceb-b94b34622384", "bridge": "br-int", "label": "tempest-ServersTestJSON-1882023778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2c6a4814c234d53bbd4844cd482fbae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0349ff40-86", "ovs_interfaceid": "0349ff40-8656-4ada-9f99-9196782155c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 959.094876] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:25:ee:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e9fa4744-8702-4973-b911-ee18192a3e4b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0349ff40-8656-4ada-9f99-9196782155c6', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 959.105312] env[62569]: DEBUG oslo.service.loopingcall [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 959.106397] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 959.106785] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-67ddab72-5525-4daa-b3e6-9cb18104f9d6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.131160] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 959.131160] env[62569]: value = "task-1250468" [ 959.131160] env[62569]: _type = "Task" [ 959.131160] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.138918] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250468, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.209146] env[62569]: DEBUG oslo_vmware.api [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250466, 'name': PowerOffVM_Task, 'duration_secs': 0.463998} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.209447] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 959.209631] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 959.209947] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1115731f-612a-4fc0-af17-3747db39ecaa {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.230297] env[62569]: DEBUG nova.compute.manager [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 959.233274] env[62569]: DEBUG nova.objects.base [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62569) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 959.435582] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250465, 'name': Rename_Task, 'duration_secs': 0.927929} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.438278] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 959.438761] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-82178fb3-bdbf-4be6-931c-c24370b92a97 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.445857] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 959.445857] env[62569]: value = "task-1250470" [ 959.445857] env[62569]: _type = "Task" [ 959.445857] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.450821] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2277d4ee-8bd5-4e95-8955-8d81caadf547 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.454758] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 959.455021] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 959.455224] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Deleting the datastore file [datastore1] cb9d8991-b7ba-4bf7-9d22-1391f4cea768 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 959.455824] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3a80dfb9-7c33-4fc9-bba8-d02abc7d4f79 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.462251] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250470, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.463423] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b82f242f-bb1c-4d01-9fc3-c0b9c4a8c227 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.467564] env[62569]: DEBUG oslo_vmware.api [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 959.467564] env[62569]: value = "task-1250471" [ 959.467564] env[62569]: _type = "Task" [ 959.467564] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.468945] env[62569]: DEBUG nova.network.neutron [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Updating instance_info_cache with network_info: [{"id": "90b3aaef-2e72-486f-be8d-a4202a6cb0c6", "address": "fa:16:3e:0c:8f:dd", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90b3aaef-2e", "ovs_interfaceid": "90b3aaef-2e72-486f-be8d-a4202a6cb0c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.501504] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00e2bb92-6c07-4f26-8eaa-1cd758b4b33f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.507245] env[62569]: DEBUG oslo_vmware.api [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250471, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.512141] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c148cb9b-d020-49e6-8cdd-0c32fc26bd05 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.527033] env[62569]: DEBUG nova.compute.provider_tree [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.590520] env[62569]: DEBUG oslo_vmware.api [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250467, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.642282] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250468, 'name': CreateVM_Task, 'duration_secs': 0.35883} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.642282] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 959.642853] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.643032] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.643391] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 959.643720] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-743b8c7a-bc06-41ec-807d-90b50aaa9730 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.648276] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 959.648276] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5213000d-6a10-b98f-0f77-07f9176bde97" [ 959.648276] env[62569]: _type = "Task" [ 959.648276] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.658508] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5213000d-6a10-b98f-0f77-07f9176bde97, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.686894] env[62569]: DEBUG oslo_concurrency.lockutils [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "09eeb937-aabc-418d-9fd3-3c7a72f4ef75-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.687146] env[62569]: DEBUG oslo_concurrency.lockutils [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "09eeb937-aabc-418d-9fd3-3c7a72f4ef75-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.687338] env[62569]: DEBUG oslo_concurrency.lockutils [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "09eeb937-aabc-418d-9fd3-3c7a72f4ef75-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.882242] env[62569]: DEBUG nova.network.neutron [req-132d2419-fd3e-428b-a7bd-57db6b7f52e8 req-92d94843-0ed0-4f2a-abc5-2827d9880723 service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Updated VIF entry in instance network info cache for port 0d27f2d0-6f98-4031-97c8-33df605f1bb0. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 959.882770] env[62569]: DEBUG nova.network.neutron [req-132d2419-fd3e-428b-a7bd-57db6b7f52e8 req-92d94843-0ed0-4f2a-abc5-2827d9880723 service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Updating instance_info_cache with network_info: [{"id": "91986023-125e-442b-b105-102630ec6036", "address": "fa:16:3e:18:83:9b", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91986023-12", "ovs_interfaceid": "91986023-125e-442b-b105-102630ec6036", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0d27f2d0-6f98-4031-97c8-33df605f1bb0", "address": "fa:16:3e:e3:fb:62", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d27f2d0-6f", "ovs_interfaceid": "0d27f2d0-6f98-4031-97c8-33df605f1bb0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.962326] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250470, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.980052] env[62569]: DEBUG oslo_vmware.api [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250471, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.243374} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.980452] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 959.980751] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 959.981057] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 959.981357] env[62569]: INFO nova.compute.manager [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Took 1.30 seconds to destroy the instance on the hypervisor. [ 959.981696] env[62569]: DEBUG oslo.service.loopingcall [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 959.981985] env[62569]: DEBUG nova.compute.manager [-] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 959.982144] env[62569]: DEBUG nova.network.neutron [-] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 960.000790] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Releasing lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.029860] env[62569]: DEBUG nova.scheduler.client.report [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 960.090201] env[62569]: DEBUG oslo_vmware.api [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250467, 'name': ReconfigVM_Task, 'duration_secs': 0.637688} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.090737] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.090914] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Reconfigured VM to attach interface {{(pid=62569) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 960.158965] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5213000d-6a10-b98f-0f77-07f9176bde97, 'name': SearchDatastore_Task, 'duration_secs': 0.016733} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.159286] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.159498] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 960.159733] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.159949] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.160081] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 960.160352] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e8441a19-ae62-43a0-83f7-364a0fc4eef2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.168623] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 960.168808] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 960.169547] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8363733e-1137-40cb-89c5-7dbc9b044520 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.174770] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 960.174770] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5215d12b-7fd6-b261-bbba-a60b426bd9cd" [ 960.174770] env[62569]: _type = "Task" [ 960.174770] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.182551] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5215d12b-7fd6-b261-bbba-a60b426bd9cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.243721] env[62569]: DEBUG nova.compute.manager [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 960.275632] env[62569]: DEBUG nova.virt.hardware [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 960.275987] env[62569]: DEBUG nova.virt.hardware [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 960.276348] env[62569]: DEBUG nova.virt.hardware [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 960.276627] env[62569]: DEBUG nova.virt.hardware [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 960.276858] env[62569]: DEBUG nova.virt.hardware [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 960.277098] env[62569]: DEBUG nova.virt.hardware [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 960.277417] env[62569]: DEBUG nova.virt.hardware [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 960.277664] env[62569]: DEBUG nova.virt.hardware [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 960.277924] env[62569]: DEBUG nova.virt.hardware [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 960.278189] env[62569]: DEBUG nova.virt.hardware [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 960.278473] env[62569]: DEBUG nova.virt.hardware [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 960.279733] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aa88676-7eca-4306-b966-312cebe2ae87 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.289909] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff718e52-34cf-4dd9-b1d0-86d061f5aeaf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.385935] env[62569]: DEBUG oslo_concurrency.lockutils [req-132d2419-fd3e-428b-a7bd-57db6b7f52e8 req-92d94843-0ed0-4f2a-abc5-2827d9880723 service nova] Releasing lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.386293] env[62569]: DEBUG nova.compute.manager [req-132d2419-fd3e-428b-a7bd-57db6b7f52e8 req-92d94843-0ed0-4f2a-abc5-2827d9880723 service nova] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Received event network-vif-plugged-0349ff40-8656-4ada-9f99-9196782155c6 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 960.387236] env[62569]: DEBUG oslo_concurrency.lockutils [req-132d2419-fd3e-428b-a7bd-57db6b7f52e8 req-92d94843-0ed0-4f2a-abc5-2827d9880723 service nova] Acquiring lock "abf56021-ddf8-4677-9e90-46dc96e89133-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.387236] env[62569]: DEBUG oslo_concurrency.lockutils [req-132d2419-fd3e-428b-a7bd-57db6b7f52e8 req-92d94843-0ed0-4f2a-abc5-2827d9880723 service nova] Lock "abf56021-ddf8-4677-9e90-46dc96e89133-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.387236] env[62569]: DEBUG oslo_concurrency.lockutils [req-132d2419-fd3e-428b-a7bd-57db6b7f52e8 req-92d94843-0ed0-4f2a-abc5-2827d9880723 service nova] Lock "abf56021-ddf8-4677-9e90-46dc96e89133-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.387236] env[62569]: DEBUG nova.compute.manager [req-132d2419-fd3e-428b-a7bd-57db6b7f52e8 req-92d94843-0ed0-4f2a-abc5-2827d9880723 service nova] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] No waiting events found dispatching network-vif-plugged-0349ff40-8656-4ada-9f99-9196782155c6 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 960.387427] env[62569]: WARNING nova.compute.manager [req-132d2419-fd3e-428b-a7bd-57db6b7f52e8 req-92d94843-0ed0-4f2a-abc5-2827d9880723 service nova] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Received unexpected event network-vif-plugged-0349ff40-8656-4ada-9f99-9196782155c6 for instance with vm_state building and task_state spawning. [ 960.387482] env[62569]: DEBUG nova.compute.manager [req-132d2419-fd3e-428b-a7bd-57db6b7f52e8 req-92d94843-0ed0-4f2a-abc5-2827d9880723 service nova] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Received event network-changed-0349ff40-8656-4ada-9f99-9196782155c6 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 960.387615] env[62569]: DEBUG nova.compute.manager [req-132d2419-fd3e-428b-a7bd-57db6b7f52e8 req-92d94843-0ed0-4f2a-abc5-2827d9880723 service nova] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Refreshing instance network info cache due to event network-changed-0349ff40-8656-4ada-9f99-9196782155c6. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 960.387801] env[62569]: DEBUG oslo_concurrency.lockutils [req-132d2419-fd3e-428b-a7bd-57db6b7f52e8 req-92d94843-0ed0-4f2a-abc5-2827d9880723 service nova] Acquiring lock "refresh_cache-abf56021-ddf8-4677-9e90-46dc96e89133" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.387936] env[62569]: DEBUG oslo_concurrency.lockutils [req-132d2419-fd3e-428b-a7bd-57db6b7f52e8 req-92d94843-0ed0-4f2a-abc5-2827d9880723 service nova] Acquired lock "refresh_cache-abf56021-ddf8-4677-9e90-46dc96e89133" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.388130] env[62569]: DEBUG nova.network.neutron [req-132d2419-fd3e-428b-a7bd-57db6b7f52e8 req-92d94843-0ed0-4f2a-abc5-2827d9880723 service nova] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Refreshing network info cache for port 0349ff40-8656-4ada-9f99-9196782155c6 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 960.459574] env[62569]: DEBUG oslo_vmware.api [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250470, 'name': PowerOnVM_Task, 'duration_secs': 0.58781} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.460034] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 960.528583] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8b742f3-94ee-4ce1-bef8-35a4f5b28a1d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.548788] env[62569]: DEBUG oslo_concurrency.lockutils [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.326s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.551362] env[62569]: DEBUG oslo_concurrency.lockutils [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.137s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.551614] env[62569]: DEBUG nova.objects.instance [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lazy-loading 'resources' on Instance uuid d1fdff70-0401-49ca-bbbb-ef5eda266c9c {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 960.552970] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e894ac99-67d8-4b65-a439-99ad35e6d7cc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.560295] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Updating instance 'b47cd2d7-0cd2-41af-8ed1-a6dfca323516' progress to 83 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 960.567711] env[62569]: DEBUG nova.compute.manager [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 960.568638] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad8cbf8b-f42d-4ac9-bd36-ef7594e105ce {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.595760] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ae05ce51-5482-49c7-93a5-7ea1ebe3bc4b tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "interface-188cac03-4034-4a02-973a-fb1906399fd1-0d27f2d0-6f98-4031-97c8-33df605f1bb0" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.760s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.685928] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5215d12b-7fd6-b261-bbba-a60b426bd9cd, 'name': SearchDatastore_Task, 'duration_secs': 0.011926} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.686808] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b46874df-0daf-4c80-8ca6-44e7dcf1729a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.692935] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 960.692935] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527b6174-3e2f-6238-0b83-ec74bf1c1272" [ 960.692935] env[62569]: _type = "Task" [ 960.692935] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.703266] env[62569]: DEBUG nova.compute.manager [req-5ade7474-7488-4c53-a1e8-4194fc4c78b0 req-02e248b8-6aa0-4685-a452-8cdabb809b3f service nova] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Received event network-vif-plugged-a274afa6-f296-447a-902c-0747076a59c9 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 960.703650] env[62569]: DEBUG oslo_concurrency.lockutils [req-5ade7474-7488-4c53-a1e8-4194fc4c78b0 req-02e248b8-6aa0-4685-a452-8cdabb809b3f service nova] Acquiring lock "bcc7f73b-083b-477e-bceb-4f984a95d219-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.703857] env[62569]: DEBUG oslo_concurrency.lockutils [req-5ade7474-7488-4c53-a1e8-4194fc4c78b0 req-02e248b8-6aa0-4685-a452-8cdabb809b3f service nova] Lock "bcc7f73b-083b-477e-bceb-4f984a95d219-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.704049] env[62569]: DEBUG oslo_concurrency.lockutils [req-5ade7474-7488-4c53-a1e8-4194fc4c78b0 req-02e248b8-6aa0-4685-a452-8cdabb809b3f service nova] Lock "bcc7f73b-083b-477e-bceb-4f984a95d219-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.704221] env[62569]: DEBUG nova.compute.manager [req-5ade7474-7488-4c53-a1e8-4194fc4c78b0 req-02e248b8-6aa0-4685-a452-8cdabb809b3f service nova] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] No waiting events found dispatching network-vif-plugged-a274afa6-f296-447a-902c-0747076a59c9 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 960.704393] env[62569]: WARNING nova.compute.manager [req-5ade7474-7488-4c53-a1e8-4194fc4c78b0 req-02e248b8-6aa0-4685-a452-8cdabb809b3f service nova] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Received unexpected event network-vif-plugged-a274afa6-f296-447a-902c-0747076a59c9 for instance with vm_state building and task_state spawning. [ 960.708267] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527b6174-3e2f-6238-0b83-ec74bf1c1272, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.727310] env[62569]: DEBUG nova.compute.manager [req-de250235-6581-47ff-8a97-ff26d3b0a8cf req-be431a2b-8e91-40dc-8d75-7e47a7a15239 service nova] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Received event network-vif-deleted-cbc4229a-f42d-463c-9a46-6a4d6b3581b0 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 960.727310] env[62569]: INFO nova.compute.manager [req-de250235-6581-47ff-8a97-ff26d3b0a8cf req-be431a2b-8e91-40dc-8d75-7e47a7a15239 service nova] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Neutron deleted interface cbc4229a-f42d-463c-9a46-6a4d6b3581b0; detaching it from the instance and deleting it from the info cache [ 960.727310] env[62569]: DEBUG nova.network.neutron [req-de250235-6581-47ff-8a97-ff26d3b0a8cf req-be431a2b-8e91-40dc-8d75-7e47a7a15239 service nova] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.739616] env[62569]: DEBUG oslo_concurrency.lockutils [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "refresh_cache-09eeb937-aabc-418d-9fd3-3c7a72f4ef75" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.739616] env[62569]: DEBUG oslo_concurrency.lockutils [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired lock "refresh_cache-09eeb937-aabc-418d-9fd3-3c7a72f4ef75" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.739616] env[62569]: DEBUG nova.network.neutron [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 960.985851] env[62569]: DEBUG nova.network.neutron [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Successfully updated port: a274afa6-f296-447a-902c-0747076a59c9 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 961.064058] env[62569]: DEBUG oslo_concurrency.lockutils [None req-452cc27a-4634-4f99-bfc3-fdcd2e34bc3d tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "af26f440-3515-4be7-9a03-8a0404c152d9" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 22.543s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.065785] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fe3b5602-a4b4-4f7a-b454-ec0b615fe240 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "af26f440-3515-4be7-9a03-8a0404c152d9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 2.966s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.066025] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fe3b5602-a4b4-4f7a-b454-ec0b615fe240 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "af26f440-3515-4be7-9a03-8a0404c152d9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.066328] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fe3b5602-a4b4-4f7a-b454-ec0b615fe240 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "af26f440-3515-4be7-9a03-8a0404c152d9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.066410] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fe3b5602-a4b4-4f7a-b454-ec0b615fe240 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "af26f440-3515-4be7-9a03-8a0404c152d9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.069280] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bba2cc1b-17bf-4034-ad17-1a55c720f0c3 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Updating instance 'b47cd2d7-0cd2-41af-8ed1-a6dfca323516' progress to 100 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 961.073225] env[62569]: INFO nova.compute.manager [None req-fe3b5602-a4b4-4f7a-b454-ec0b615fe240 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Terminating instance [ 961.082850] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2f6d83cc-dc82-48a1-b310-c9c9d0f87682 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "b6e5eefc-8c06-445b-a3af-9404578b6179" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 32.413s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.130499] env[62569]: DEBUG nova.compute.manager [req-b0b92fac-572e-4e6e-9636-3c28289ed5d4 req-ad2bf5d7-a4d9-4387-b3f7-3578fd7fec81 service nova] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Received event network-changed-a274afa6-f296-447a-902c-0747076a59c9 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 961.130568] env[62569]: DEBUG nova.compute.manager [req-b0b92fac-572e-4e6e-9636-3c28289ed5d4 req-ad2bf5d7-a4d9-4387-b3f7-3578fd7fec81 service nova] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Refreshing instance network info cache due to event network-changed-a274afa6-f296-447a-902c-0747076a59c9. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 961.130729] env[62569]: DEBUG oslo_concurrency.lockutils [req-b0b92fac-572e-4e6e-9636-3c28289ed5d4 req-ad2bf5d7-a4d9-4387-b3f7-3578fd7fec81 service nova] Acquiring lock "refresh_cache-bcc7f73b-083b-477e-bceb-4f984a95d219" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.130870] env[62569]: DEBUG oslo_concurrency.lockutils [req-b0b92fac-572e-4e6e-9636-3c28289ed5d4 req-ad2bf5d7-a4d9-4387-b3f7-3578fd7fec81 service nova] Acquired lock "refresh_cache-bcc7f73b-083b-477e-bceb-4f984a95d219" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.131450] env[62569]: DEBUG nova.network.neutron [req-b0b92fac-572e-4e6e-9636-3c28289ed5d4 req-ad2bf5d7-a4d9-4387-b3f7-3578fd7fec81 service nova] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Refreshing network info cache for port a274afa6-f296-447a-902c-0747076a59c9 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 961.153754] env[62569]: DEBUG nova.network.neutron [req-132d2419-fd3e-428b-a7bd-57db6b7f52e8 req-92d94843-0ed0-4f2a-abc5-2827d9880723 service nova] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Updated VIF entry in instance network info cache for port 0349ff40-8656-4ada-9f99-9196782155c6. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 961.154159] env[62569]: DEBUG nova.network.neutron [req-132d2419-fd3e-428b-a7bd-57db6b7f52e8 req-92d94843-0ed0-4f2a-abc5-2827d9880723 service nova] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Updating instance_info_cache with network_info: [{"id": "0349ff40-8656-4ada-9f99-9196782155c6", "address": "fa:16:3e:25:ee:d5", "network": {"id": "884d3248-07f1-4cf7-8ceb-b94b34622384", "bridge": "br-int", "label": "tempest-ServersTestJSON-1882023778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2c6a4814c234d53bbd4844cd482fbae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0349ff40-86", "ovs_interfaceid": "0349ff40-8656-4ada-9f99-9196782155c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.156828] env[62569]: DEBUG nova.network.neutron [-] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.207583] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527b6174-3e2f-6238-0b83-ec74bf1c1272, 'name': SearchDatastore_Task, 'duration_secs': 0.014132} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.207583] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.207583] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] abf56021-ddf8-4677-9e90-46dc96e89133/abf56021-ddf8-4677-9e90-46dc96e89133.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 961.207583] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3127684a-f8c8-4a5a-bb23-6da186de0d8e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.216814] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 961.216814] env[62569]: value = "task-1250472" [ 961.216814] env[62569]: _type = "Task" [ 961.216814] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.222565] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250472, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.232831] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5ab92fe6-9eb4-4305-9a1a-9e99db7f58b8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.241989] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49f48471-d685-459f-af73-c080dcff8165 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.273159] env[62569]: DEBUG nova.compute.manager [req-de250235-6581-47ff-8a97-ff26d3b0a8cf req-be431a2b-8e91-40dc-8d75-7e47a7a15239 service nova] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Detach interface failed, port_id=cbc4229a-f42d-463c-9a46-6a4d6b3581b0, reason: Instance cb9d8991-b7ba-4bf7-9d22-1391f4cea768 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 961.301138] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f527a6d-311b-44bf-ace7-453701e18c35 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.309061] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b7cb85a-a1cb-4378-b3ed-dca4b65a91a2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.343754] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b70cdfb-fb34-4f1a-b8de-82ea65345fe9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.353860] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae11612a-7f79-49c0-b970-53ad8fca3ea1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.375501] env[62569]: DEBUG nova.compute.provider_tree [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.492904] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "refresh_cache-bcc7f73b-083b-477e-bceb-4f984a95d219" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.578369] env[62569]: DEBUG nova.compute.manager [None req-fe3b5602-a4b4-4f7a-b454-ec0b615fe240 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 961.578678] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-fe3b5602-a4b4-4f7a-b454-ec0b615fe240 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 961.579055] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-29a83de7-8555-4ff9-9d9b-cb1f2450600a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.590147] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5823fb6-3836-4a00-9bd4-ef6f675be143 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.622549] env[62569]: WARNING nova.virt.vmwareapi.vmops [None req-fe3b5602-a4b4-4f7a-b454-ec0b615fe240 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance af26f440-3515-4be7-9a03-8a0404c152d9 could not be found. [ 961.622814] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-fe3b5602-a4b4-4f7a-b454-ec0b615fe240 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 961.623017] env[62569]: INFO nova.compute.manager [None req-fe3b5602-a4b4-4f7a-b454-ec0b615fe240 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 961.623288] env[62569]: DEBUG oslo.service.loopingcall [None req-fe3b5602-a4b4-4f7a-b454-ec0b615fe240 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.623570] env[62569]: DEBUG nova.compute.manager [-] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 961.623673] env[62569]: DEBUG nova.network.neutron [-] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 961.657311] env[62569]: DEBUG oslo_concurrency.lockutils [req-132d2419-fd3e-428b-a7bd-57db6b7f52e8 req-92d94843-0ed0-4f2a-abc5-2827d9880723 service nova] Releasing lock "refresh_cache-abf56021-ddf8-4677-9e90-46dc96e89133" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.659040] env[62569]: INFO nova.compute.manager [-] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Took 1.68 seconds to deallocate network for instance. [ 961.706990] env[62569]: DEBUG nova.network.neutron [req-b0b92fac-572e-4e6e-9636-3c28289ed5d4 req-ad2bf5d7-a4d9-4387-b3f7-3578fd7fec81 service nova] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 961.725016] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250472, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471346} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.725289] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] abf56021-ddf8-4677-9e90-46dc96e89133/abf56021-ddf8-4677-9e90-46dc96e89133.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 961.725518] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 961.725825] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ddc02d9c-f07c-4e0a-933e-ccd301dbb988 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.732811] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 961.732811] env[62569]: value = "task-1250473" [ 961.732811] env[62569]: _type = "Task" [ 961.732811] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.744382] env[62569]: DEBUG nova.network.neutron [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Updating instance_info_cache with network_info: [{"id": "c967324f-ab14-480b-a8b8-710db5827863", "address": "fa:16:3e:f9:c6:da", "network": {"id": "c61b7867-7897-43ae-8f5b-19834c4390ea", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-512066189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ba7e96d226942bd99ab29fe703c6fea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc967324f-ab", "ovs_interfaceid": "c967324f-ab14-480b-a8b8-710db5827863", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.747756] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250473, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.842082] env[62569]: DEBUG nova.network.neutron [req-b0b92fac-572e-4e6e-9636-3c28289ed5d4 req-ad2bf5d7-a4d9-4387-b3f7-3578fd7fec81 service nova] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.850365] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "interface-188cac03-4034-4a02-973a-fb1906399fd1-0d27f2d0-6f98-4031-97c8-33df605f1bb0" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.850641] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "interface-188cac03-4034-4a02-973a-fb1906399fd1-0d27f2d0-6f98-4031-97c8-33df605f1bb0" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.878423] env[62569]: DEBUG nova.scheduler.client.report [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 962.167732] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.242812] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250473, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.277234} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.244506] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 962.245424] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c471e3c-9d5c-44cb-b81a-c21463963567 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.248095] env[62569]: DEBUG oslo_concurrency.lockutils [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Releasing lock "refresh_cache-09eeb937-aabc-418d-9fd3-3c7a72f4ef75" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.273239] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] abf56021-ddf8-4677-9e90-46dc96e89133/abf56021-ddf8-4677-9e90-46dc96e89133.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 962.273578] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-076f5377-99b9-4d65-8796-4069204cc4f8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.299454] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 962.299454] env[62569]: value = "task-1250474" [ 962.299454] env[62569]: _type = "Task" [ 962.299454] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.307165] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250474, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.345151] env[62569]: DEBUG oslo_concurrency.lockutils [req-b0b92fac-572e-4e6e-9636-3c28289ed5d4 req-ad2bf5d7-a4d9-4387-b3f7-3578fd7fec81 service nova] Releasing lock "refresh_cache-bcc7f73b-083b-477e-bceb-4f984a95d219" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.345568] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "refresh_cache-bcc7f73b-083b-477e-bceb-4f984a95d219" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.345919] env[62569]: DEBUG nova.network.neutron [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 962.353418] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.353609] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.354487] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3be9f07-ebc3-4fd3-a499-0cfd2395685e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.374403] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ed6493b-24a3-4caf-834d-6511dbf74b6d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.395272] env[62569]: DEBUG oslo_concurrency.lockutils [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.844s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.403564] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Reconfiguring VM to detach interface {{(pid=62569) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 962.404185] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.237s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.404449] env[62569]: DEBUG nova.objects.instance [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lazy-loading 'resources' on Instance uuid cb9d8991-b7ba-4bf7-9d22-1391f4cea768 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 962.406248] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2416706-0245-417f-a769-dac2c92315e5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.431173] env[62569]: DEBUG oslo_vmware.api [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 962.431173] env[62569]: value = "task-1250475" [ 962.431173] env[62569]: _type = "Task" [ 962.431173] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.434826] env[62569]: INFO nova.scheduler.client.report [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Deleted allocations for instance d1fdff70-0401-49ca-bbbb-ef5eda266c9c [ 962.447059] env[62569]: DEBUG oslo_vmware.api [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250475, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.477106] env[62569]: DEBUG nova.network.neutron [-] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.772023] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da166861-df41-4f48-b0df-296bb0ee15d2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.791560] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3d906a1-4e20-4128-bc74-c61a619a96ab {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.795053] env[62569]: DEBUG oslo_concurrency.lockutils [None req-970f75a6-baab-4a66-a42e-6ae5487fb87b tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "b47cd2d7-0cd2-41af-8ed1-a6dfca323516" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.795053] env[62569]: DEBUG oslo_concurrency.lockutils [None req-970f75a6-baab-4a66-a42e-6ae5487fb87b tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "b47cd2d7-0cd2-41af-8ed1-a6dfca323516" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.795235] env[62569]: DEBUG nova.compute.manager [None req-970f75a6-baab-4a66-a42e-6ae5487fb87b tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Going to confirm migration 2 {{(pid=62569) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5156}} [ 962.804024] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Updating instance '09eeb937-aabc-418d-9fd3-3c7a72f4ef75' progress to 83 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 962.812204] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250474, 'name': ReconfigVM_Task, 'duration_secs': 0.285831} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.812513] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Reconfigured VM instance instance-0000005f to attach disk [datastore2] abf56021-ddf8-4677-9e90-46dc96e89133/abf56021-ddf8-4677-9e90-46dc96e89133.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 962.813228] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cb7dd3b3-e32b-4e7d-8e3a-68696df07d37 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.818980] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 962.818980] env[62569]: value = "task-1250476" [ 962.818980] env[62569]: _type = "Task" [ 962.818980] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.827080] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250476, 'name': Rename_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.879165] env[62569]: DEBUG nova.network.neutron [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 962.940201] env[62569]: DEBUG oslo_vmware.api [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250475, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.947469] env[62569]: DEBUG oslo_concurrency.lockutils [None req-294d77a3-7ca0-4192-aa1e-aa0dd6d0fce7 tempest-ImagesTestJSON-1070393917 tempest-ImagesTestJSON-1070393917-project-member] Lock "d1fdff70-0401-49ca-bbbb-ef5eda266c9c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.434s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.980335] env[62569]: INFO nova.compute.manager [-] [instance: af26f440-3515-4be7-9a03-8a0404c152d9] Took 1.36 seconds to deallocate network for instance. [ 963.027845] env[62569]: DEBUG nova.network.neutron [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Updating instance_info_cache with network_info: [{"id": "a274afa6-f296-447a-902c-0747076a59c9", "address": "fa:16:3e:ef:60:b7", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa274afa6-f2", "ovs_interfaceid": "a274afa6-f296-447a-902c-0747076a59c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.103693] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f533330e-0567-4d77-8edf-9fb688b0a5bf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.111847] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52186255-45bd-4443-803e-1ac771555df6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.145293] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfee7577-23b4-4a48-b0cb-398c2b0186ac {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.148506] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3d25b7a-d066-4d95-866f-741729ad0060 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.154892] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bb98cab1-4b0f-44d5-99ff-089256a40138 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Suspending the VM {{(pid=62569) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 963.157058] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-8b9d217a-4c4d-4fdb-991e-f32dac8096f1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.159536] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4eed49d-a6a7-4620-95ef-3144053bade1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.173602] env[62569]: DEBUG nova.compute.provider_tree [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 963.177561] env[62569]: DEBUG oslo_vmware.api [None req-bb98cab1-4b0f-44d5-99ff-089256a40138 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 963.177561] env[62569]: value = "task-1250477" [ 963.177561] env[62569]: _type = "Task" [ 963.177561] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.185473] env[62569]: DEBUG oslo_vmware.api [None req-bb98cab1-4b0f-44d5-99ff-089256a40138 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250477, 'name': SuspendVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.310422] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 963.310985] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-27ac9343-ba36-4fa3-aa8d-6db4d40ea061 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.321177] env[62569]: DEBUG oslo_vmware.api [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 963.321177] env[62569]: value = "task-1250478" [ 963.321177] env[62569]: _type = "Task" [ 963.321177] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.332646] env[62569]: DEBUG oslo_vmware.api [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250478, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.335886] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250476, 'name': Rename_Task, 'duration_secs': 0.156487} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.336338] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 963.336459] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-af7d8ffb-496e-40eb-b0f7-a32520453ec9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.341893] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 963.341893] env[62569]: value = "task-1250479" [ 963.341893] env[62569]: _type = "Task" [ 963.341893] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.353499] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250479, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.362147] env[62569]: DEBUG oslo_concurrency.lockutils [None req-970f75a6-baab-4a66-a42e-6ae5487fb87b tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.362354] env[62569]: DEBUG oslo_concurrency.lockutils [None req-970f75a6-baab-4a66-a42e-6ae5487fb87b tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.362537] env[62569]: DEBUG nova.network.neutron [None req-970f75a6-baab-4a66-a42e-6ae5487fb87b tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 963.362727] env[62569]: DEBUG nova.objects.instance [None req-970f75a6-baab-4a66-a42e-6ae5487fb87b tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lazy-loading 'info_cache' on Instance uuid b47cd2d7-0cd2-41af-8ed1-a6dfca323516 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 963.440965] env[62569]: DEBUG oslo_vmware.api [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250475, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.531678] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "refresh_cache-bcc7f73b-083b-477e-bceb-4f984a95d219" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.531868] env[62569]: DEBUG nova.compute.manager [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Instance network_info: |[{"id": "a274afa6-f296-447a-902c-0747076a59c9", "address": "fa:16:3e:ef:60:b7", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa274afa6-f2", "ovs_interfaceid": "a274afa6-f296-447a-902c-0747076a59c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 963.532289] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ef:60:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73915082-a1b0-460b-b24d-97588fc9cb29', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a274afa6-f296-447a-902c-0747076a59c9', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 963.540573] env[62569]: DEBUG oslo.service.loopingcall [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 963.540838] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 963.544455] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-101d5d33-6a1e-4d52-a08b-88c0b5144452 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.577480] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 963.577480] env[62569]: value = "task-1250480" [ 963.577480] env[62569]: _type = "Task" [ 963.577480] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.585939] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250480, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.689976] env[62569]: DEBUG oslo_vmware.api [None req-bb98cab1-4b0f-44d5-99ff-089256a40138 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250477, 'name': SuspendVM_Task} progress is 66%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.708315] env[62569]: DEBUG nova.scheduler.client.report [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Updated inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with generation 126 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 963.708673] env[62569]: DEBUG nova.compute.provider_tree [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Updating resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 generation from 126 to 127 during operation: update_inventory {{(pid=62569) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 963.708900] env[62569]: DEBUG nova.compute.provider_tree [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 963.834890] env[62569]: DEBUG oslo_vmware.api [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250478, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.854080] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250479, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.941493] env[62569]: DEBUG oslo_vmware.api [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250475, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.013446] env[62569]: DEBUG oslo_concurrency.lockutils [None req-fe3b5602-a4b4-4f7a-b454-ec0b615fe240 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "af26f440-3515-4be7-9a03-8a0404c152d9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.948s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.087148] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250480, 'name': CreateVM_Task, 'duration_secs': 0.359514} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.087337] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 964.088074] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.088276] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.088600] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 964.088870] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b71771b-6289-4da4-91a3-f48cd8c0091e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.093526] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 964.093526] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52e31024-f66e-7e2c-dde7-4980774f9a95" [ 964.093526] env[62569]: _type = "Task" [ 964.093526] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.101266] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52e31024-f66e-7e2c-dde7-4980774f9a95, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.188960] env[62569]: DEBUG oslo_vmware.api [None req-bb98cab1-4b0f-44d5-99ff-089256a40138 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250477, 'name': SuspendVM_Task, 'duration_secs': 0.783372} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.189531] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bb98cab1-4b0f-44d5-99ff-089256a40138 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Suspended the VM {{(pid=62569) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 964.189775] env[62569]: DEBUG nova.compute.manager [None req-bb98cab1-4b0f-44d5-99ff-089256a40138 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 964.190687] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a244ce1-8cd7-4659-9f0e-726ca800bb54 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.214665] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.810s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.246393] env[62569]: INFO nova.scheduler.client.report [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Deleted allocations for instance cb9d8991-b7ba-4bf7-9d22-1391f4cea768 [ 964.342302] env[62569]: DEBUG oslo_vmware.api [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250478, 'name': PowerOnVM_Task, 'duration_secs': 1.006858} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.342302] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 964.342302] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-25481e87-e074-4359-bbc4-b76c61615846 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Updating instance '09eeb937-aabc-418d-9fd3-3c7a72f4ef75' progress to 100 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 964.352785] env[62569]: DEBUG oslo_vmware.api [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250479, 'name': PowerOnVM_Task, 'duration_secs': 0.578202} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.353065] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 964.353376] env[62569]: INFO nova.compute.manager [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Took 7.47 seconds to spawn the instance on the hypervisor. [ 964.353745] env[62569]: DEBUG nova.compute.manager [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 964.354637] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a00a30fe-2652-42cf-91d0-dd02c9f97ad3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.442448] env[62569]: DEBUG oslo_vmware.api [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250475, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.608474] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52e31024-f66e-7e2c-dde7-4980774f9a95, 'name': SearchDatastore_Task, 'duration_secs': 0.010609} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.609825] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.609825] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 964.609825] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.609825] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.609825] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 964.610146] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-803828a5-d841-4cd7-963e-3947d5114319 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.628743] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 964.629085] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 964.629811] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b61b1b4e-0fc7-4b9b-ab0c-db361053609b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.635631] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 964.635631] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528e1d06-f132-04a0-6121-e6cdeb805677" [ 964.635631] env[62569]: _type = "Task" [ 964.635631] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.643496] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528e1d06-f132-04a0-6121-e6cdeb805677, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.644428] env[62569]: DEBUG nova.network.neutron [None req-970f75a6-baab-4a66-a42e-6ae5487fb87b tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Updating instance_info_cache with network_info: [{"id": "90b3aaef-2e72-486f-be8d-a4202a6cb0c6", "address": "fa:16:3e:0c:8f:dd", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90b3aaef-2e", "ovs_interfaceid": "90b3aaef-2e72-486f-be8d-a4202a6cb0c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.758706] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0d98fd3f-6e37-4a9c-9064-b0a6e559e00b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cb9d8991-b7ba-4bf7-9d22-1391f4cea768" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.585s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.872474] env[62569]: INFO nova.compute.manager [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Took 24.49 seconds to build instance. [ 964.942571] env[62569]: DEBUG oslo_vmware.api [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250475, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.970754] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "35350164-e295-40ea-85a6-e8e79c630e0f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.970996] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "35350164-e295-40ea-85a6-e8e79c630e0f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.146883] env[62569]: DEBUG oslo_concurrency.lockutils [None req-970f75a6-baab-4a66-a42e-6ae5487fb87b tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Releasing lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.146981] env[62569]: DEBUG nova.objects.instance [None req-970f75a6-baab-4a66-a42e-6ae5487fb87b tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lazy-loading 'migration_context' on Instance uuid b47cd2d7-0cd2-41af-8ed1-a6dfca323516 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.149425] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528e1d06-f132-04a0-6121-e6cdeb805677, 'name': SearchDatastore_Task, 'duration_secs': 0.057043} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.149425] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-918aee84-4d0d-4147-a2c7-5b6e835e2ea8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.154696] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 965.154696] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52bfbfcd-d343-8424-06c8-5a3169ef4cfe" [ 965.154696] env[62569]: _type = "Task" [ 965.154696] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.163354] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52bfbfcd-d343-8424-06c8-5a3169ef4cfe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.374689] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a7e8f046-cd08-4435-a78d-1517db867900 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "abf56021-ddf8-4677-9e90-46dc96e89133" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.443369] env[62569]: DEBUG oslo_vmware.api [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250475, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.473685] env[62569]: DEBUG nova.compute.manager [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 965.656577] env[62569]: DEBUG nova.objects.base [None req-970f75a6-baab-4a66-a42e-6ae5487fb87b tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62569) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 965.657694] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-265ff49d-af78-4cec-8bc2-3043b6bd7ab0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.684648] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-facdf4e2-f627-4fab-ab59-be6ee7154b22 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.687110] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52bfbfcd-d343-8424-06c8-5a3169ef4cfe, 'name': SearchDatastore_Task, 'duration_secs': 0.026724} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.687338] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.687707] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] bcc7f73b-083b-477e-bceb-4f984a95d219/bcc7f73b-083b-477e-bceb-4f984a95d219.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 965.688262] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-87b772ac-b74d-4f6c-abc8-423377c4b6f9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.691645] env[62569]: DEBUG oslo_vmware.api [None req-970f75a6-baab-4a66-a42e-6ae5487fb87b tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 965.691645] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a055c9-9e73-f621-334f-df9ebe22f8e7" [ 965.691645] env[62569]: _type = "Task" [ 965.691645] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.696615] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 965.696615] env[62569]: value = "task-1250481" [ 965.696615] env[62569]: _type = "Task" [ 965.696615] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.699620] env[62569]: DEBUG oslo_vmware.api [None req-970f75a6-baab-4a66-a42e-6ae5487fb87b tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a055c9-9e73-f621-334f-df9ebe22f8e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.708168] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250481, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.729758] env[62569]: INFO nova.compute.manager [None req-4e4b6da8-c811-4f14-8cd6-27e23c33fb68 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Resuming [ 965.730418] env[62569]: DEBUG nova.objects.instance [None req-4e4b6da8-c811-4f14-8cd6-27e23c33fb68 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lazy-loading 'flavor' on Instance uuid b6e5eefc-8c06-445b-a3af-9404578b6179 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.944824] env[62569]: DEBUG oslo_vmware.api [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250475, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.008017] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.008017] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.008698] env[62569]: INFO nova.compute.claims [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 966.204405] env[62569]: DEBUG oslo_vmware.api [None req-970f75a6-baab-4a66-a42e-6ae5487fb87b tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a055c9-9e73-f621-334f-df9ebe22f8e7, 'name': SearchDatastore_Task, 'duration_secs': 0.007787} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.208842] env[62569]: DEBUG oslo_concurrency.lockutils [None req-970f75a6-baab-4a66-a42e-6ae5487fb87b tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.216145] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "abf56021-ddf8-4677-9e90-46dc96e89133" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.216607] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "abf56021-ddf8-4677-9e90-46dc96e89133" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.216991] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "abf56021-ddf8-4677-9e90-46dc96e89133-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.217336] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "abf56021-ddf8-4677-9e90-46dc96e89133-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.217788] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "abf56021-ddf8-4677-9e90-46dc96e89133-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.220176] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250481, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.224020] env[62569]: INFO nova.compute.manager [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Terminating instance [ 966.224020] env[62569]: DEBUG oslo_concurrency.lockutils [None req-262450e9-79ae-4558-88f2-16108de5835a tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "09eeb937-aabc-418d-9fd3-3c7a72f4ef75" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.224020] env[62569]: DEBUG oslo_concurrency.lockutils [None req-262450e9-79ae-4558-88f2-16108de5835a tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "09eeb937-aabc-418d-9fd3-3c7a72f4ef75" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.224020] env[62569]: DEBUG nova.compute.manager [None req-262450e9-79ae-4558-88f2-16108de5835a tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Going to confirm migration 3 {{(pid=62569) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5156}} [ 966.444032] env[62569]: DEBUG oslo_vmware.api [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250475, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.713824] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250481, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.653846} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.714293] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] bcc7f73b-083b-477e-bceb-4f984a95d219/bcc7f73b-083b-477e-bceb-4f984a95d219.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 966.714694] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 966.715469] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-55b66eea-14e1-437e-a0e7-08e90e66f4ba {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.723959] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 966.723959] env[62569]: value = "task-1250482" [ 966.723959] env[62569]: _type = "Task" [ 966.723959] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.730829] env[62569]: DEBUG nova.compute.manager [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 966.731324] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 966.733987] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2380f414-370c-49e7-b6a7-096dcc3045f5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.747045] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250482, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.749680] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 966.750210] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c0eb825-4f28-47d2-9419-a149355bde6b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.756582] env[62569]: DEBUG oslo_vmware.api [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 966.756582] env[62569]: value = "task-1250483" [ 966.756582] env[62569]: _type = "Task" [ 966.756582] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.765340] env[62569]: DEBUG oslo_vmware.api [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250483, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.865855] env[62569]: DEBUG oslo_concurrency.lockutils [None req-262450e9-79ae-4558-88f2-16108de5835a tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "refresh_cache-09eeb937-aabc-418d-9fd3-3c7a72f4ef75" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.866065] env[62569]: DEBUG oslo_concurrency.lockutils [None req-262450e9-79ae-4558-88f2-16108de5835a tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired lock "refresh_cache-09eeb937-aabc-418d-9fd3-3c7a72f4ef75" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.866614] env[62569]: DEBUG nova.network.neutron [None req-262450e9-79ae-4558-88f2-16108de5835a tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 966.866889] env[62569]: DEBUG nova.objects.instance [None req-262450e9-79ae-4558-88f2-16108de5835a tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lazy-loading 'info_cache' on Instance uuid 09eeb937-aabc-418d-9fd3-3c7a72f4ef75 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 966.944730] env[62569]: DEBUG oslo_vmware.api [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250475, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.220383] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fecb86e-2d4d-4257-b89b-85ae7372cb21 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.230538] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7841562-0ade-4c78-afd0-2b69c1adf102 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.236737] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250482, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.21074} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.241024] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 967.241024] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c537fd20-2fb6-4fa0-b075-5e8bedcb8ef5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.267276] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4e4b6da8-c811-4f14-8cd6-27e23c33fb68 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.267546] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4e4b6da8-c811-4f14-8cd6-27e23c33fb68 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquired lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.269204] env[62569]: DEBUG nova.network.neutron [None req-4e4b6da8-c811-4f14-8cd6-27e23c33fb68 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 967.277212] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd5e8115-29d7-496b-9664-c0eb5d03b4c2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.298186] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] bcc7f73b-083b-477e-bceb-4f984a95d219/bcc7f73b-083b-477e-bceb-4f984a95d219.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 967.299597] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4a6c40d-d93c-432c-8c2a-b46b5cfe83d2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.319938] env[62569]: DEBUG oslo_vmware.api [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250483, 'name': PowerOffVM_Task, 'duration_secs': 0.272468} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.322262] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17d31ab3-150e-4452-b213-88902ccbdb7c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.325529] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 967.325714] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 967.329024] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-35513368-a53f-410c-b00a-e84d26bec194 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.329024] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 967.329024] env[62569]: value = "task-1250484" [ 967.329024] env[62569]: _type = "Task" [ 967.329024] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.341086] env[62569]: DEBUG nova.compute.provider_tree [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.346667] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250484, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.397070] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 967.397315] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 967.397556] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Deleting the datastore file [datastore2] abf56021-ddf8-4677-9e90-46dc96e89133 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 967.397832] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-295ace08-8a35-4293-9805-59a22da7c1bd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.405254] env[62569]: DEBUG oslo_vmware.api [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 967.405254] env[62569]: value = "task-1250486" [ 967.405254] env[62569]: _type = "Task" [ 967.405254] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.413550] env[62569]: DEBUG oslo_vmware.api [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250486, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.446079] env[62569]: DEBUG oslo_vmware.api [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250475, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.537705] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "d56faf39-6caf-4489-98d7-342fc5fb7d40" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.537705] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "d56faf39-6caf-4489-98d7-342fc5fb7d40" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.842068] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250484, 'name': ReconfigVM_Task, 'duration_secs': 0.300511} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.842387] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Reconfigured VM instance instance-00000060 to attach disk [datastore2] bcc7f73b-083b-477e-bceb-4f984a95d219/bcc7f73b-083b-477e-bceb-4f984a95d219.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 967.843528] env[62569]: DEBUG nova.scheduler.client.report [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 967.846896] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-61fe82e4-34c6-47d3-b1d3-26d675464d8e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.854100] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 967.854100] env[62569]: value = "task-1250487" [ 967.854100] env[62569]: _type = "Task" [ 967.854100] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.869316] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250487, 'name': Rename_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.915951] env[62569]: DEBUG oslo_vmware.api [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250486, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.500415} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.916377] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 967.916641] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 967.916808] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 967.916994] env[62569]: INFO nova.compute.manager [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Took 1.19 seconds to destroy the instance on the hypervisor. [ 967.917250] env[62569]: DEBUG oslo.service.loopingcall [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 967.917640] env[62569]: DEBUG nova.compute.manager [-] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 967.917640] env[62569]: DEBUG nova.network.neutron [-] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 967.953618] env[62569]: DEBUG oslo_vmware.api [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250475, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.042642] env[62569]: DEBUG nova.compute.manager [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 968.190567] env[62569]: DEBUG nova.network.neutron [None req-4e4b6da8-c811-4f14-8cd6-27e23c33fb68 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Updating instance_info_cache with network_info: [{"id": "cfddbae7-9ab0-4bc6-aea1-46a0cde743fc", "address": "fa:16:3e:b7:d6:80", "network": {"id": "406dd658-8168-4e73-9b97-861929bfaa2e", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-693938985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "de8605118d744a93bce54e897aa849b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "257e5ea7-8b80-4301-9900-a754f1fe2031", "external-id": "nsx-vlan-transportzone-682", "segmentation_id": 682, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfddbae7-9a", "ovs_interfaceid": "cfddbae7-9ab0-4bc6-aea1-46a0cde743fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.281407] env[62569]: DEBUG nova.compute.manager [req-3039992e-e191-4c48-a681-92fc657d0945 req-8d3c33dc-fc4f-4e4f-a2a5-d2d8ba3d1224 service nova] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Received event network-vif-deleted-0349ff40-8656-4ada-9f99-9196782155c6 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 968.281644] env[62569]: INFO nova.compute.manager [req-3039992e-e191-4c48-a681-92fc657d0945 req-8d3c33dc-fc4f-4e4f-a2a5-d2d8ba3d1224 service nova] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Neutron deleted interface 0349ff40-8656-4ada-9f99-9196782155c6; detaching it from the instance and deleting it from the info cache [ 968.281885] env[62569]: DEBUG nova.network.neutron [req-3039992e-e191-4c48-a681-92fc657d0945 req-8d3c33dc-fc4f-4e4f-a2a5-d2d8ba3d1224 service nova] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.329113] env[62569]: DEBUG nova.network.neutron [None req-262450e9-79ae-4558-88f2-16108de5835a tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Updating instance_info_cache with network_info: [{"id": "c967324f-ab14-480b-a8b8-710db5827863", "address": "fa:16:3e:f9:c6:da", "network": {"id": "c61b7867-7897-43ae-8f5b-19834c4390ea", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-512066189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ba7e96d226942bd99ab29fe703c6fea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc967324f-ab", "ovs_interfaceid": "c967324f-ab14-480b-a8b8-710db5827863", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.350176] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.343s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.351167] env[62569]: DEBUG nova.compute.manager [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 968.353701] env[62569]: DEBUG oslo_concurrency.lockutils [None req-970f75a6-baab-4a66-a42e-6ae5487fb87b tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 2.145s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.376446] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250487, 'name': Rename_Task, 'duration_secs': 0.139966} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.376643] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 968.376898] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8fd43c6b-177c-46c0-b1c5-7a9628c37e03 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.384254] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 968.384254] env[62569]: value = "task-1250488" [ 968.384254] env[62569]: _type = "Task" [ 968.384254] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.392990] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250488, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.423733] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c4a8626b-2969-42e9-b440-ebbfc5417e5d tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.424027] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c4a8626b-2969-42e9-b440-ebbfc5417e5d tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.449801] env[62569]: DEBUG oslo_vmware.api [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250475, 'name': ReconfigVM_Task, 'duration_secs': 5.775129} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.450113] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.453065] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Reconfigured VM to detach interface {{(pid=62569) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 968.567336] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.676267] env[62569]: DEBUG nova.network.neutron [-] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.693962] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4e4b6da8-c811-4f14-8cd6-27e23c33fb68 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Releasing lock "refresh_cache-b6e5eefc-8c06-445b-a3af-9404578b6179" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.695520] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ca7a749-3762-4a08-9366-497dc7a7c87d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.709535] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4e4b6da8-c811-4f14-8cd6-27e23c33fb68 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Resuming the VM {{(pid=62569) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 968.710342] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a0f7e890-a7b8-440d-8e1c-74dbac604b99 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.716490] env[62569]: DEBUG oslo_vmware.api [None req-4e4b6da8-c811-4f14-8cd6-27e23c33fb68 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 968.716490] env[62569]: value = "task-1250489" [ 968.716490] env[62569]: _type = "Task" [ 968.716490] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.724921] env[62569]: DEBUG oslo_vmware.api [None req-4e4b6da8-c811-4f14-8cd6-27e23c33fb68 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250489, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.788631] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-05e3109c-ebe4-4bcd-b65d-1ba55e580a2b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.798760] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88171ccb-63bf-4707-8971-e38d970a7be4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.833514] env[62569]: DEBUG oslo_concurrency.lockutils [None req-262450e9-79ae-4558-88f2-16108de5835a tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Releasing lock "refresh_cache-09eeb937-aabc-418d-9fd3-3c7a72f4ef75" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.833825] env[62569]: DEBUG nova.objects.instance [None req-262450e9-79ae-4558-88f2-16108de5835a tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lazy-loading 'migration_context' on Instance uuid 09eeb937-aabc-418d-9fd3-3c7a72f4ef75 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 968.840433] env[62569]: DEBUG nova.compute.manager [req-3039992e-e191-4c48-a681-92fc657d0945 req-8d3c33dc-fc4f-4e4f-a2a5-d2d8ba3d1224 service nova] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Detach interface failed, port_id=0349ff40-8656-4ada-9f99-9196782155c6, reason: Instance abf56021-ddf8-4677-9e90-46dc96e89133 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 968.864397] env[62569]: DEBUG nova.compute.utils [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 968.866427] env[62569]: DEBUG nova.compute.manager [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 968.866427] env[62569]: DEBUG nova.network.neutron [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 968.899018] env[62569]: DEBUG oslo_vmware.api [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250488, 'name': PowerOnVM_Task, 'duration_secs': 0.462423} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.899018] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 968.899018] env[62569]: INFO nova.compute.manager [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Took 8.65 seconds to spawn the instance on the hypervisor. [ 968.899018] env[62569]: DEBUG nova.compute.manager [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 968.899018] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d30043cf-384d-40b7-9af8-f1949c3411b2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.929193] env[62569]: DEBUG nova.compute.utils [None req-c4a8626b-2969-42e9-b440-ebbfc5417e5d tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 968.933701] env[62569]: DEBUG nova.policy [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5da4f6b7a6784a73bd3fed04b275041b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bdba6022e3d4697a336ca28ca4eccec', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 969.142050] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f30ae1f-3d9a-4529-a11f-f594b4eb8294 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.151170] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d14ecda6-1096-4055-9167-13ec928be6bf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.182019] env[62569]: INFO nova.compute.manager [-] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Took 1.26 seconds to deallocate network for instance. [ 969.183768] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35bbf1d1-052a-457c-bfd0-8423617e13c6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.194992] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ecd48ff-4771-404c-bfc5-e7398f346b6f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.209124] env[62569]: DEBUG nova.compute.provider_tree [None req-970f75a6-baab-4a66-a42e-6ae5487fb87b tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 969.226510] env[62569]: DEBUG oslo_vmware.api [None req-4e4b6da8-c811-4f14-8cd6-27e23c33fb68 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250489, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.244399] env[62569]: DEBUG nova.network.neutron [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Successfully created port: 95f1fe41-3f4f-44ce-adb6-7639e7268c5d {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 969.343032] env[62569]: DEBUG nova.objects.base [None req-262450e9-79ae-4558-88f2-16108de5835a tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Object Instance<09eeb937-aabc-418d-9fd3-3c7a72f4ef75> lazy-loaded attributes: info_cache,migration_context {{(pid=62569) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 969.344089] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d98d6c68-ceed-4002-a869-af0d5f7f9442 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.364949] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f2f4b3a-d17b-4fe3-b6a8-610cc6249ea5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.367369] env[62569]: DEBUG nova.compute.manager [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 969.373679] env[62569]: DEBUG oslo_vmware.api [None req-262450e9-79ae-4558-88f2-16108de5835a tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 969.373679] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52afdba8-9fb2-e8e4-0869-281c34958ce3" [ 969.373679] env[62569]: _type = "Task" [ 969.373679] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.381789] env[62569]: DEBUG oslo_vmware.api [None req-262450e9-79ae-4558-88f2-16108de5835a tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52afdba8-9fb2-e8e4-0869-281c34958ce3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.430750] env[62569]: INFO nova.compute.manager [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Took 14.57 seconds to build instance. [ 969.435876] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c4a8626b-2969-42e9-b440-ebbfc5417e5d tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.012s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.692206] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.714840] env[62569]: DEBUG nova.scheduler.client.report [None req-970f75a6-baab-4a66-a42e-6ae5487fb87b tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 969.732041] env[62569]: DEBUG oslo_vmware.api [None req-4e4b6da8-c811-4f14-8cd6-27e23c33fb68 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250489, 'name': PowerOnVM_Task, 'duration_secs': 0.86614} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.735019] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4e4b6da8-c811-4f14-8cd6-27e23c33fb68 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Resumed the VM {{(pid=62569) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 969.735019] env[62569]: DEBUG nova.compute.manager [None req-4e4b6da8-c811-4f14-8cd6-27e23c33fb68 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 969.735019] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b157c10b-6726-4ede-bd46-17756f35a31b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.894500] env[62569]: DEBUG oslo_vmware.api [None req-262450e9-79ae-4558-88f2-16108de5835a tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52afdba8-9fb2-e8e4-0869-281c34958ce3, 'name': SearchDatastore_Task, 'duration_secs': 0.008224} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.894500] env[62569]: DEBUG oslo_concurrency.lockutils [None req-262450e9-79ae-4558-88f2-16108de5835a tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.933221] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc8ae922-ac14-4489-aa1e-602a634eb217 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "bcc7f73b-083b-477e-bceb-4f984a95d219" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.110s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.993424] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.994212] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.994212] env[62569]: DEBUG nova.network.neutron [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 970.054393] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquiring lock "9b5a1451-af46-4cd8-8e7a-f4560d8499e6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.054393] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Lock "9b5a1451-af46-4cd8-8e7a-f4560d8499e6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.383315] env[62569]: DEBUG nova.compute.manager [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 970.428174] env[62569]: DEBUG nova.virt.hardware [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 970.428269] env[62569]: DEBUG nova.virt.hardware [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 970.428462] env[62569]: DEBUG nova.virt.hardware [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 970.428805] env[62569]: DEBUG nova.virt.hardware [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 970.428859] env[62569]: DEBUG nova.virt.hardware [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 970.429467] env[62569]: DEBUG nova.virt.hardware [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 970.429467] env[62569]: DEBUG nova.virt.hardware [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 970.429467] env[62569]: DEBUG nova.virt.hardware [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 970.429612] env[62569]: DEBUG nova.virt.hardware [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 970.429739] env[62569]: DEBUG nova.virt.hardware [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 970.429918] env[62569]: DEBUG nova.virt.hardware [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 970.431607] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95b1fd03-acea-45e9-925c-ed266dbf8083 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.440453] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d79f964e-f9a6-4963-8643-078a77be6727 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.536055] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c4a8626b-2969-42e9-b440-ebbfc5417e5d tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.536055] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c4a8626b-2969-42e9-b440-ebbfc5417e5d tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.536055] env[62569]: INFO nova.compute.manager [None req-c4a8626b-2969-42e9-b440-ebbfc5417e5d tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Attaching volume 4494e8a8-22d4-4334-a304-ace87e30e8ed to /dev/sdb [ 970.560116] env[62569]: DEBUG nova.compute.manager [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 970.593437] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96329769-ffe7-4bc4-ab06-d15395c82ed2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.605818] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03792c8f-f57a-490e-bcd8-6687b5a15664 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.621280] env[62569]: DEBUG nova.virt.block_device [None req-c4a8626b-2969-42e9-b440-ebbfc5417e5d tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Updating existing volume attachment record: 2a78764e-9900-4e28-893b-181b9ec9881b {{(pid=62569) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 970.724838] env[62569]: DEBUG oslo_concurrency.lockutils [None req-970f75a6-baab-4a66-a42e-6ae5487fb87b tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.371s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.725136] env[62569]: DEBUG nova.compute.manager [None req-970f75a6-baab-4a66-a42e-6ae5487fb87b tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=62569) _confirm_resize /opt/stack/nova/nova/compute/manager.py:5282}} [ 970.733400] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.167s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.735604] env[62569]: INFO nova.compute.claims [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 970.981053] env[62569]: INFO nova.compute.manager [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Rebuilding instance [ 971.016670] env[62569]: DEBUG nova.compute.manager [req-1ea1b6c9-9a4f-4bd0-8aec-30a801ff3eea req-231c4ea1-d441-4e0d-9348-fbfa427dcb8c service nova] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Received event network-vif-plugged-95f1fe41-3f4f-44ce-adb6-7639e7268c5d {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 971.016871] env[62569]: DEBUG oslo_concurrency.lockutils [req-1ea1b6c9-9a4f-4bd0-8aec-30a801ff3eea req-231c4ea1-d441-4e0d-9348-fbfa427dcb8c service nova] Acquiring lock "35350164-e295-40ea-85a6-e8e79c630e0f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.017031] env[62569]: DEBUG oslo_concurrency.lockutils [req-1ea1b6c9-9a4f-4bd0-8aec-30a801ff3eea req-231c4ea1-d441-4e0d-9348-fbfa427dcb8c service nova] Lock "35350164-e295-40ea-85a6-e8e79c630e0f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.018847] env[62569]: DEBUG oslo_concurrency.lockutils [req-1ea1b6c9-9a4f-4bd0-8aec-30a801ff3eea req-231c4ea1-d441-4e0d-9348-fbfa427dcb8c service nova] Lock "35350164-e295-40ea-85a6-e8e79c630e0f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.019204] env[62569]: DEBUG nova.compute.manager [req-1ea1b6c9-9a4f-4bd0-8aec-30a801ff3eea req-231c4ea1-d441-4e0d-9348-fbfa427dcb8c service nova] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] No waiting events found dispatching network-vif-plugged-95f1fe41-3f4f-44ce-adb6-7639e7268c5d {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 971.019414] env[62569]: WARNING nova.compute.manager [req-1ea1b6c9-9a4f-4bd0-8aec-30a801ff3eea req-231c4ea1-d441-4e0d-9348-fbfa427dcb8c service nova] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Received unexpected event network-vif-plugged-95f1fe41-3f4f-44ce-adb6-7639e7268c5d for instance with vm_state building and task_state spawning. [ 971.031096] env[62569]: INFO nova.network.neutron [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Port 0d27f2d0-6f98-4031-97c8-33df605f1bb0 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 971.031096] env[62569]: DEBUG nova.network.neutron [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Updating instance_info_cache with network_info: [{"id": "91986023-125e-442b-b105-102630ec6036", "address": "fa:16:3e:18:83:9b", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91986023-12", "ovs_interfaceid": "91986023-125e-442b-b105-102630ec6036", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.049561] env[62569]: DEBUG nova.compute.manager [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 971.050477] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e19761d5-41aa-4392-9978-1f8e04a6f406 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.084252] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.191183] env[62569]: DEBUG nova.network.neutron [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Successfully updated port: 95f1fe41-3f4f-44ce-adb6-7639e7268c5d {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 971.305276] env[62569]: INFO nova.scheduler.client.report [None req-970f75a6-baab-4a66-a42e-6ae5487fb87b tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Deleted allocation for migration 54fd4982-eb4e-4229-a3e0-21648ce18eb7 [ 971.531729] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.695327] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "refresh_cache-35350164-e295-40ea-85a6-e8e79c630e0f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.695327] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired lock "refresh_cache-35350164-e295-40ea-85a6-e8e79c630e0f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.695327] env[62569]: DEBUG nova.network.neutron [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 971.814124] env[62569]: DEBUG oslo_concurrency.lockutils [None req-970f75a6-baab-4a66-a42e-6ae5487fb87b tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "b47cd2d7-0cd2-41af-8ed1-a6dfca323516" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.019s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.947794] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6153330c-5687-4b59-a748-46cc545ac6e2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.954214] env[62569]: DEBUG nova.objects.instance [None req-73ab93ca-705f-42a2-b4ef-0196562a1ed6 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lazy-loading 'flavor' on Instance uuid b47cd2d7-0cd2-41af-8ed1-a6dfca323516 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 971.958594] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8f12472-f4e0-469d-b83d-867187aa6b14 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.999639] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e77011c-7b1c-4866-8ab9-1eed55a6ef00 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.008666] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cca13769-9a35-45bc-a388-35e4a59666f6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.022423] env[62569]: DEBUG nova.compute.provider_tree [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.035395] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cfc11a7e-45f4-48e7-b0b7-d1c4b96356a2 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "interface-188cac03-4034-4a02-973a-fb1906399fd1-0d27f2d0-6f98-4031-97c8-33df605f1bb0" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.185s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.065955] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 972.066285] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-485c03b2-458b-47f5-8132-bd4c1bb6ae31 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.074261] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 972.074261] env[62569]: value = "task-1250491" [ 972.074261] env[62569]: _type = "Task" [ 972.074261] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.081836] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250491, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.187787] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "interface-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e-0d27f2d0-6f98-4031-97c8-33df605f1bb0" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.188070] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "interface-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e-0d27f2d0-6f98-4031-97c8-33df605f1bb0" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.188453] env[62569]: DEBUG nova.objects.instance [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lazy-loading 'flavor' on Instance uuid 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 972.226764] env[62569]: DEBUG nova.network.neutron [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 972.372409] env[62569]: DEBUG nova.network.neutron [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Updating instance_info_cache with network_info: [{"id": "95f1fe41-3f4f-44ce-adb6-7639e7268c5d", "address": "fa:16:3e:a3:b6:08", "network": {"id": "30d3bf4a-8f55-492f-a049-181e29731818", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1244949475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bdba6022e3d4697a336ca28ca4eccec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95f1fe41-3f", "ovs_interfaceid": "95f1fe41-3f4f-44ce-adb6-7639e7268c5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.462860] env[62569]: DEBUG oslo_concurrency.lockutils [None req-73ab93ca-705f-42a2-b4ef-0196562a1ed6 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.462860] env[62569]: DEBUG oslo_concurrency.lockutils [None req-73ab93ca-705f-42a2-b4ef-0196562a1ed6 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.462860] env[62569]: DEBUG nova.network.neutron [None req-73ab93ca-705f-42a2-b4ef-0196562a1ed6 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 972.462860] env[62569]: DEBUG nova.objects.instance [None req-73ab93ca-705f-42a2-b4ef-0196562a1ed6 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lazy-loading 'info_cache' on Instance uuid b47cd2d7-0cd2-41af-8ed1-a6dfca323516 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 972.528022] env[62569]: DEBUG nova.scheduler.client.report [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 972.584284] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250491, 'name': PowerOffVM_Task, 'duration_secs': 0.202344} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.584571] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 972.584804] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 972.585631] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ef2a877-6711-4b6c-814a-f0925c050eb2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.592167] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 972.592426] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5487c495-1fc6-4dbf-ae7c-843e15d6c88a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.655829] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 972.656066] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 972.656261] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Deleting the datastore file [datastore2] bcc7f73b-083b-477e-bceb-4f984a95d219 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 972.656533] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1a080981-d1ac-4158-8bfe-e904cfc9d4ec {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.663014] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 972.663014] env[62569]: value = "task-1250493" [ 972.663014] env[62569]: _type = "Task" [ 972.663014] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.670160] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250493, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.783125] env[62569]: DEBUG nova.objects.instance [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lazy-loading 'pci_requests' on Instance uuid 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 972.875449] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Releasing lock "refresh_cache-35350164-e295-40ea-85a6-e8e79c630e0f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.875820] env[62569]: DEBUG nova.compute.manager [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Instance network_info: |[{"id": "95f1fe41-3f4f-44ce-adb6-7639e7268c5d", "address": "fa:16:3e:a3:b6:08", "network": {"id": "30d3bf4a-8f55-492f-a049-181e29731818", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1244949475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bdba6022e3d4697a336ca28ca4eccec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95f1fe41-3f", "ovs_interfaceid": "95f1fe41-3f4f-44ce-adb6-7639e7268c5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 972.875915] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a3:b6:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '983826cf-6390-4ec6-bf97-30a1060947fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '95f1fe41-3f4f-44ce-adb6-7639e7268c5d', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 972.883555] env[62569]: DEBUG oslo.service.loopingcall [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 972.883807] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 972.884042] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e3293df9-02c3-4e4a-9ab0-f84ad2e1cc94 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.903489] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 972.903489] env[62569]: value = "task-1250495" [ 972.903489] env[62569]: _type = "Task" [ 972.903489] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.910926] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250495, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.965062] env[62569]: DEBUG nova.objects.base [None req-73ab93ca-705f-42a2-b4ef-0196562a1ed6 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62569) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 973.031457] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.298s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.032522] env[62569]: DEBUG nova.compute.manager [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 973.036035] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.343s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.036035] env[62569]: DEBUG nova.objects.instance [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lazy-loading 'resources' on Instance uuid abf56021-ddf8-4677-9e90-46dc96e89133 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.174150] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250493, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.193553} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.174150] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 973.174150] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 973.174150] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 973.202444] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.202701] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.236215] env[62569]: DEBUG nova.compute.manager [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Received event network-changed-95f1fe41-3f4f-44ce-adb6-7639e7268c5d {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 973.236436] env[62569]: DEBUG nova.compute.manager [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Refreshing instance network info cache due to event network-changed-95f1fe41-3f4f-44ce-adb6-7639e7268c5d. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 973.236742] env[62569]: DEBUG oslo_concurrency.lockutils [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] Acquiring lock "refresh_cache-35350164-e295-40ea-85a6-e8e79c630e0f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.237770] env[62569]: DEBUG oslo_concurrency.lockutils [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] Acquired lock "refresh_cache-35350164-e295-40ea-85a6-e8e79c630e0f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.237770] env[62569]: DEBUG nova.network.neutron [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Refreshing network info cache for port 95f1fe41-3f4f-44ce-adb6-7639e7268c5d {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 973.286101] env[62569]: DEBUG nova.objects.base [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Object Instance<93eb1a09-269a-44c8-8f9d-3a2b7e0d715e> lazy-loaded attributes: flavor,pci_requests {{(pid=62569) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 973.286324] env[62569]: DEBUG nova.network.neutron [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 973.413968] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250495, 'name': CreateVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.538519] env[62569]: DEBUG nova.compute.utils [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 973.539960] env[62569]: DEBUG nova.compute.manager [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 973.540150] env[62569]: DEBUG nova.network.neutron [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 973.565954] env[62569]: DEBUG nova.policy [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ab6d20862c54432cbafdda33ccc974ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '530ca7e7924743ab91a362a064a3111b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 973.628668] env[62569]: DEBUG nova.policy [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '50780a0da3b14122aa6287c1a482f671', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7f7fb3e698324f8e851aefbee74ac669', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 973.717564] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.718867] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Starting heal instance info cache {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 973.763132] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e07d626-a51f-4248-8c6b-16c5f845c916 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.771369] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7d71806-cff4-4f03-a983-e7a86ae1ab72 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.802897] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db7ef800-b4c8-4322-b973-4eed0d5a352c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.811421] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3df5757a-d8cc-4b76-b8a5-1db8575bf486 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.824349] env[62569]: DEBUG nova.compute.provider_tree [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 973.860652] env[62569]: DEBUG nova.network.neutron [None req-73ab93ca-705f-42a2-b4ef-0196562a1ed6 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Updating instance_info_cache with network_info: [{"id": "90b3aaef-2e72-486f-be8d-a4202a6cb0c6", "address": "fa:16:3e:0c:8f:dd", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90b3aaef-2e", "ovs_interfaceid": "90b3aaef-2e72-486f-be8d-a4202a6cb0c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.917312] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250495, 'name': CreateVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.044810] env[62569]: DEBUG nova.compute.manager [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 974.178226] env[62569]: DEBUG nova.network.neutron [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Updated VIF entry in instance network info cache for port 95f1fe41-3f4f-44ce-adb6-7639e7268c5d. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 974.178976] env[62569]: DEBUG nova.network.neutron [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Updating instance_info_cache with network_info: [{"id": "95f1fe41-3f4f-44ce-adb6-7639e7268c5d", "address": "fa:16:3e:a3:b6:08", "network": {"id": "30d3bf4a-8f55-492f-a049-181e29731818", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1244949475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bdba6022e3d4697a336ca28ca4eccec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95f1fe41-3f", "ovs_interfaceid": "95f1fe41-3f4f-44ce-adb6-7639e7268c5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.232778] env[62569]: DEBUG nova.virt.hardware [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 974.233138] env[62569]: DEBUG nova.virt.hardware [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 974.233918] env[62569]: DEBUG nova.virt.hardware [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 974.233918] env[62569]: DEBUG nova.virt.hardware [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 974.233918] env[62569]: DEBUG nova.virt.hardware [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 974.233918] env[62569]: DEBUG nova.virt.hardware [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 974.234264] env[62569]: DEBUG nova.virt.hardware [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 974.234264] env[62569]: DEBUG nova.virt.hardware [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 974.235033] env[62569]: DEBUG nova.virt.hardware [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 974.235033] env[62569]: DEBUG nova.virt.hardware [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 974.235033] env[62569]: DEBUG nova.virt.hardware [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 974.236051] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01bbc2b6-e71c-48a5-a762-ca19230d2c43 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.245038] env[62569]: DEBUG nova.network.neutron [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Successfully created port: b6625a0e-7282-4ade-b7f0-3e5374ec8e64 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 974.254010] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e42ed0df-eead-4cf0-adba-e900a2f8c995 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.269558] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ef:60:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73915082-a1b0-460b-b24d-97588fc9cb29', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a274afa6-f296-447a-902c-0747076a59c9', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 974.277594] env[62569]: DEBUG oslo.service.loopingcall [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 974.277905] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 974.278169] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f68a3b78-1423-4eb2-a5b2-57604bf2d0d4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.297316] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 974.297316] env[62569]: value = "task-1250496" [ 974.297316] env[62569]: _type = "Task" [ 974.297316] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.304716] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250496, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.346612] env[62569]: ERROR nova.scheduler.client.report [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [req-e823bc63-81fd-4d15-b235-936ca4dda468] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID fa06556a-5785-4014-b8bd-bc240a0cf716. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e823bc63-81fd-4d15-b235-936ca4dda468"}]} [ 974.363752] env[62569]: DEBUG oslo_concurrency.lockutils [None req-73ab93ca-705f-42a2-b4ef-0196562a1ed6 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Releasing lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.366425] env[62569]: DEBUG nova.scheduler.client.report [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Refreshing inventories for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 974.381652] env[62569]: DEBUG nova.scheduler.client.report [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Updating ProviderTree inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 974.381892] env[62569]: DEBUG nova.compute.provider_tree [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 974.395358] env[62569]: DEBUG nova.scheduler.client.report [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Refreshing aggregate associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, aggregates: None {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 974.414518] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250495, 'name': CreateVM_Task, 'duration_secs': 1.326355} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.415407] env[62569]: DEBUG nova.scheduler.client.report [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Refreshing trait associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 974.417290] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 974.418275] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.418438] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.418753] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 974.419054] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9da0fddb-ace3-4e0a-84c6-af0f10f8dd54 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.423527] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 974.423527] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520185ec-d246-a9b3-ca09-aa8c9cd2e12c" [ 974.423527] env[62569]: _type = "Task" [ 974.423527] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.431327] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520185ec-d246-a9b3-ca09-aa8c9cd2e12c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.623984] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4a0b4de-c833-4afc-9a2e-ac144298df83 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.632122] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c90d5ee-e4ac-4112-bb13-76a1a1b2b39c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.666598] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1599de6d-d6bf-4145-80c5-a525718f71b0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.674986] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86946499-732b-49aa-b209-f68f0e248e85 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.681545] env[62569]: DEBUG oslo_concurrency.lockutils [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] Releasing lock "refresh_cache-35350164-e295-40ea-85a6-e8e79c630e0f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.681848] env[62569]: DEBUG nova.compute.manager [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Received event network-changed-91986023-125e-442b-b105-102630ec6036 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 974.682042] env[62569]: DEBUG nova.compute.manager [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Refreshing instance network info cache due to event network-changed-91986023-125e-442b-b105-102630ec6036. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 974.682281] env[62569]: DEBUG oslo_concurrency.lockutils [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] Acquiring lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.682447] env[62569]: DEBUG oslo_concurrency.lockutils [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] Acquired lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.682620] env[62569]: DEBUG nova.network.neutron [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Refreshing network info cache for port 91986023-125e-442b-b105-102630ec6036 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 974.691618] env[62569]: DEBUG nova.compute.provider_tree [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 974.807915] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250496, 'name': CreateVM_Task, 'duration_secs': 0.302375} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.808794] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 974.808794] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.934413] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520185ec-d246-a9b3-ca09-aa8c9cd2e12c, 'name': SearchDatastore_Task, 'duration_secs': 0.011342} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.934729] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.935443] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 974.935443] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.935443] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.935628] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 974.935913] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.936241] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 974.936483] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d284918b-0b94-4414-8c3d-c094875d7383 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.938329] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d996afe-1862-4860-ba11-82764955f89c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.943594] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 974.943594] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52cb013b-f8f8-a52e-90fe-b62b7fb3558a" [ 974.943594] env[62569]: _type = "Task" [ 974.943594] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.948039] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 974.948231] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 974.949280] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3997e43-5219-4e23-a12c-2ee958a25389 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.954489] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52cb013b-f8f8-a52e-90fe-b62b7fb3558a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.957440] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 974.957440] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52eda084-606a-2532-5ed6-2a0427cdfb8d" [ 974.957440] env[62569]: _type = "Task" [ 974.957440] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.965280] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52eda084-606a-2532-5ed6-2a0427cdfb8d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.055589] env[62569]: DEBUG nova.compute.manager [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 975.080856] env[62569]: DEBUG nova.virt.hardware [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 975.081193] env[62569]: DEBUG nova.virt.hardware [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 975.081416] env[62569]: DEBUG nova.virt.hardware [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 975.081642] env[62569]: DEBUG nova.virt.hardware [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 975.081829] env[62569]: DEBUG nova.virt.hardware [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 975.081991] env[62569]: DEBUG nova.virt.hardware [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 975.082237] env[62569]: DEBUG nova.virt.hardware [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 975.082426] env[62569]: DEBUG nova.virt.hardware [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 975.082603] env[62569]: DEBUG nova.virt.hardware [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 975.082798] env[62569]: DEBUG nova.virt.hardware [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 975.082979] env[62569]: DEBUG nova.virt.hardware [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 975.083861] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-470f671c-7cf0-4314-b6be-3b11fd9e2e34 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.091648] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b26e59c7-f383-4ae3-ab0e-5f23ad396bdd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.178545] env[62569]: DEBUG nova.network.neutron [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Successfully updated port: 0d27f2d0-6f98-4031-97c8-33df605f1bb0 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 975.180572] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4a8626b-2969-42e9-b440-ebbfc5417e5d tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Volume attach. Driver type: vmdk {{(pid=62569) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 975.180815] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4a8626b-2969-42e9-b440-ebbfc5417e5d tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269488', 'volume_id': '4494e8a8-22d4-4334-a304-ace87e30e8ed', 'name': 'volume-4494e8a8-22d4-4334-a304-ace87e30e8ed', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '98089a34-074b-4bdb-92ae-f9e23d2551f1', 'attached_at': '', 'detached_at': '', 'volume_id': '4494e8a8-22d4-4334-a304-ace87e30e8ed', 'serial': '4494e8a8-22d4-4334-a304-ace87e30e8ed'} {{(pid=62569) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 975.181823] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b7d294f-859d-4b3c-98f9-c9d9c134daaf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.203414] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-247e4b5a-7dbb-49d1-b19b-026829c53885 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.229072] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4a8626b-2969-42e9-b440-ebbfc5417e5d tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] volume-4494e8a8-22d4-4334-a304-ace87e30e8ed/volume-4494e8a8-22d4-4334-a304-ace87e30e8ed.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 975.230595] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cfcce4db-2622-4d0d-8681-b5c62b759fa6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.244191] env[62569]: DEBUG nova.scheduler.client.report [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Updated inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with generation 131 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 975.244436] env[62569]: DEBUG nova.compute.provider_tree [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Updating resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 generation from 131 to 132 during operation: update_inventory {{(pid=62569) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 975.244619] env[62569]: DEBUG nova.compute.provider_tree [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 975.257048] env[62569]: DEBUG oslo_vmware.api [None req-c4a8626b-2969-42e9-b440-ebbfc5417e5d tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 975.257048] env[62569]: value = "task-1250497" [ 975.257048] env[62569]: _type = "Task" [ 975.257048] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.264099] env[62569]: DEBUG nova.compute.manager [req-ab9fedfc-1146-4c8d-af46-04a0e11b9235 req-6ee82fab-535d-4627-be1b-49249c08169f service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Received event network-vif-plugged-0d27f2d0-6f98-4031-97c8-33df605f1bb0 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 975.264317] env[62569]: DEBUG oslo_concurrency.lockutils [req-ab9fedfc-1146-4c8d-af46-04a0e11b9235 req-6ee82fab-535d-4627-be1b-49249c08169f service nova] Acquiring lock "93eb1a09-269a-44c8-8f9d-3a2b7e0d715e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.264601] env[62569]: DEBUG oslo_concurrency.lockutils [req-ab9fedfc-1146-4c8d-af46-04a0e11b9235 req-6ee82fab-535d-4627-be1b-49249c08169f service nova] Lock "93eb1a09-269a-44c8-8f9d-3a2b7e0d715e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.264733] env[62569]: DEBUG oslo_concurrency.lockutils [req-ab9fedfc-1146-4c8d-af46-04a0e11b9235 req-6ee82fab-535d-4627-be1b-49249c08169f service nova] Lock "93eb1a09-269a-44c8-8f9d-3a2b7e0d715e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.264837] env[62569]: DEBUG nova.compute.manager [req-ab9fedfc-1146-4c8d-af46-04a0e11b9235 req-6ee82fab-535d-4627-be1b-49249c08169f service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] No waiting events found dispatching network-vif-plugged-0d27f2d0-6f98-4031-97c8-33df605f1bb0 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 975.264997] env[62569]: WARNING nova.compute.manager [req-ab9fedfc-1146-4c8d-af46-04a0e11b9235 req-6ee82fab-535d-4627-be1b-49249c08169f service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Received unexpected event network-vif-plugged-0d27f2d0-6f98-4031-97c8-33df605f1bb0 for instance with vm_state active and task_state None. [ 975.265176] env[62569]: DEBUG nova.compute.manager [req-ab9fedfc-1146-4c8d-af46-04a0e11b9235 req-6ee82fab-535d-4627-be1b-49249c08169f service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Received event network-changed-0d27f2d0-6f98-4031-97c8-33df605f1bb0 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 975.265328] env[62569]: DEBUG nova.compute.manager [req-ab9fedfc-1146-4c8d-af46-04a0e11b9235 req-6ee82fab-535d-4627-be1b-49249c08169f service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Refreshing instance network info cache due to event network-changed-0d27f2d0-6f98-4031-97c8-33df605f1bb0. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 975.265505] env[62569]: DEBUG oslo_concurrency.lockutils [req-ab9fedfc-1146-4c8d-af46-04a0e11b9235 req-6ee82fab-535d-4627-be1b-49249c08169f service nova] Acquiring lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.265644] env[62569]: DEBUG oslo_concurrency.lockutils [req-ab9fedfc-1146-4c8d-af46-04a0e11b9235 req-6ee82fab-535d-4627-be1b-49249c08169f service nova] Acquired lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.265940] env[62569]: DEBUG nova.network.neutron [req-ab9fedfc-1146-4c8d-af46-04a0e11b9235 req-6ee82fab-535d-4627-be1b-49249c08169f service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Refreshing network info cache for port 0d27f2d0-6f98-4031-97c8-33df605f1bb0 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 975.270507] env[62569]: DEBUG oslo_vmware.api [None req-c4a8626b-2969-42e9-b440-ebbfc5417e5d tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250497, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.371983] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-73ab93ca-705f-42a2-b4ef-0196562a1ed6 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 975.372350] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e2159c87-9542-412a-a4ba-9f07aa6965c0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.379907] env[62569]: DEBUG oslo_vmware.api [None req-73ab93ca-705f-42a2-b4ef-0196562a1ed6 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 975.379907] env[62569]: value = "task-1250498" [ 975.379907] env[62569]: _type = "Task" [ 975.379907] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.390739] env[62569]: DEBUG oslo_vmware.api [None req-73ab93ca-705f-42a2-b4ef-0196562a1ed6 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250498, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.426172] env[62569]: DEBUG nova.network.neutron [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Updated VIF entry in instance network info cache for port 91986023-125e-442b-b105-102630ec6036. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 975.426557] env[62569]: DEBUG nova.network.neutron [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Updating instance_info_cache with network_info: [{"id": "91986023-125e-442b-b105-102630ec6036", "address": "fa:16:3e:18:83:9b", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91986023-12", "ovs_interfaceid": "91986023-125e-442b-b105-102630ec6036", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.455664] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52cb013b-f8f8-a52e-90fe-b62b7fb3558a, 'name': SearchDatastore_Task, 'duration_secs': 0.019299} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.456155] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.456359] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 975.456619] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.466522] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52eda084-606a-2532-5ed6-2a0427cdfb8d, 'name': SearchDatastore_Task, 'duration_secs': 0.00881} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.467336] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7351ddc-acf3-4141-a691-e53d3037beb7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.472473] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 975.472473] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f33ce4-1636-00d9-6f47-ac0f789c9db2" [ 975.472473] env[62569]: _type = "Task" [ 975.472473] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.479913] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f33ce4-1636-00d9-6f47-ac0f789c9db2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.685944] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.751903] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.716s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.754472] env[62569]: DEBUG oslo_concurrency.lockutils [None req-262450e9-79ae-4558-88f2-16108de5835a tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 5.862s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.778725] env[62569]: DEBUG oslo_vmware.api [None req-c4a8626b-2969-42e9-b440-ebbfc5417e5d tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250497, 'name': ReconfigVM_Task, 'duration_secs': 0.385663} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.779222] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4a8626b-2969-42e9-b440-ebbfc5417e5d tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Reconfigured VM instance instance-0000005a to attach disk [datastore2] volume-4494e8a8-22d4-4334-a304-ace87e30e8ed/volume-4494e8a8-22d4-4334-a304-ace87e30e8ed.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 975.784862] env[62569]: INFO nova.scheduler.client.report [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Deleted allocations for instance abf56021-ddf8-4677-9e90-46dc96e89133 [ 975.792186] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7114469-08c2-4d6b-a3c4-3f3a3a9b0a8c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.811548] env[62569]: DEBUG oslo_vmware.api [None req-c4a8626b-2969-42e9-b440-ebbfc5417e5d tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 975.811548] env[62569]: value = "task-1250499" [ 975.811548] env[62569]: _type = "Task" [ 975.811548] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.822300] env[62569]: DEBUG oslo_vmware.api [None req-c4a8626b-2969-42e9-b440-ebbfc5417e5d tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250499, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.890029] env[62569]: DEBUG oslo_vmware.api [None req-73ab93ca-705f-42a2-b4ef-0196562a1ed6 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250498, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.929029] env[62569]: DEBUG oslo_concurrency.lockutils [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] Releasing lock "refresh_cache-188cac03-4034-4a02-973a-fb1906399fd1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.929327] env[62569]: DEBUG nova.compute.manager [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Received event network-changed-6ca4b28a-1255-4f88-a7b6-54b0378768b1 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 975.929504] env[62569]: DEBUG nova.compute.manager [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Refreshing instance network info cache due to event network-changed-6ca4b28a-1255-4f88-a7b6-54b0378768b1. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 975.929698] env[62569]: DEBUG oslo_concurrency.lockutils [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] Acquiring lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.986040] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f33ce4-1636-00d9-6f47-ac0f789c9db2, 'name': SearchDatastore_Task, 'duration_secs': 0.022417} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.986406] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.986583] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 35350164-e295-40ea-85a6-e8e79c630e0f/35350164-e295-40ea-85a6-e8e79c630e0f.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 975.986880] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.987085] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 975.987311] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1d350704-32b7-4edb-8212-ae70800f73ff {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.989729] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ba030d9d-06c4-4f25-bf58-31c94e941b9c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.996761] env[62569]: DEBUG nova.network.neutron [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Successfully updated port: b6625a0e-7282-4ade-b7f0-3e5374ec8e64 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 975.999692] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 975.999692] env[62569]: value = "task-1250500" [ 975.999692] env[62569]: _type = "Task" [ 975.999692] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.005285] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 976.006122] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 976.007597] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8e330be-f167-40f2-908b-f6ae9f863703 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.018139] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250500, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.021172] env[62569]: DEBUG nova.compute.manager [req-5f2782e5-dd79-43f4-b99b-2e35f9847712 req-c6511048-49ee-449a-b71d-79cb37d98dda service nova] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Received event network-vif-plugged-b6625a0e-7282-4ade-b7f0-3e5374ec8e64 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 976.021618] env[62569]: DEBUG oslo_concurrency.lockutils [req-5f2782e5-dd79-43f4-b99b-2e35f9847712 req-c6511048-49ee-449a-b71d-79cb37d98dda service nova] Acquiring lock "d56faf39-6caf-4489-98d7-342fc5fb7d40-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.021910] env[62569]: DEBUG oslo_concurrency.lockutils [req-5f2782e5-dd79-43f4-b99b-2e35f9847712 req-c6511048-49ee-449a-b71d-79cb37d98dda service nova] Lock "d56faf39-6caf-4489-98d7-342fc5fb7d40-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.022165] env[62569]: DEBUG oslo_concurrency.lockutils [req-5f2782e5-dd79-43f4-b99b-2e35f9847712 req-c6511048-49ee-449a-b71d-79cb37d98dda service nova] Lock "d56faf39-6caf-4489-98d7-342fc5fb7d40-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.022451] env[62569]: DEBUG nova.compute.manager [req-5f2782e5-dd79-43f4-b99b-2e35f9847712 req-c6511048-49ee-449a-b71d-79cb37d98dda service nova] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] No waiting events found dispatching network-vif-plugged-b6625a0e-7282-4ade-b7f0-3e5374ec8e64 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 976.022674] env[62569]: WARNING nova.compute.manager [req-5f2782e5-dd79-43f4-b99b-2e35f9847712 req-c6511048-49ee-449a-b71d-79cb37d98dda service nova] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Received unexpected event network-vif-plugged-b6625a0e-7282-4ade-b7f0-3e5374ec8e64 for instance with vm_state building and task_state spawning. [ 976.024940] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 976.024940] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]522fe663-30b1-6c64-b93b-1bb78c4f7899" [ 976.024940] env[62569]: _type = "Task" [ 976.024940] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.028809] env[62569]: DEBUG nova.network.neutron [req-ab9fedfc-1146-4c8d-af46-04a0e11b9235 req-6ee82fab-535d-4627-be1b-49249c08169f service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Added VIF to instance network info cache for port 0d27f2d0-6f98-4031-97c8-33df605f1bb0. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3546}} [ 976.029277] env[62569]: DEBUG nova.network.neutron [req-ab9fedfc-1146-4c8d-af46-04a0e11b9235 req-6ee82fab-535d-4627-be1b-49249c08169f service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Updating instance_info_cache with network_info: [{"id": "6ca4b28a-1255-4f88-a7b6-54b0378768b1", "address": "fa:16:3e:03:e7:ad", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ca4b28a-12", "ovs_interfaceid": "6ca4b28a-1255-4f88-a7b6-54b0378768b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0d27f2d0-6f98-4031-97c8-33df605f1bb0", "address": "fa:16:3e:e3:fb:62", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d27f2d0-6f", "ovs_interfaceid": "0d27f2d0-6f98-4031-97c8-33df605f1bb0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.037960] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]522fe663-30b1-6c64-b93b-1bb78c4f7899, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.295241] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f7471493-0bff-458b-b431-7d9214e3999c tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "abf56021-ddf8-4677-9e90-46dc96e89133" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.079s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.323769] env[62569]: DEBUG oslo_vmware.api [None req-c4a8626b-2969-42e9-b440-ebbfc5417e5d tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250499, 'name': ReconfigVM_Task, 'duration_secs': 0.153551} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.324098] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4a8626b-2969-42e9-b440-ebbfc5417e5d tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269488', 'volume_id': '4494e8a8-22d4-4334-a304-ace87e30e8ed', 'name': 'volume-4494e8a8-22d4-4334-a304-ace87e30e8ed', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '98089a34-074b-4bdb-92ae-f9e23d2551f1', 'attached_at': '', 'detached_at': '', 'volume_id': '4494e8a8-22d4-4334-a304-ace87e30e8ed', 'serial': '4494e8a8-22d4-4334-a304-ace87e30e8ed'} {{(pid=62569) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 976.391836] env[62569]: DEBUG oslo_vmware.api [None req-73ab93ca-705f-42a2-b4ef-0196562a1ed6 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250498, 'name': PowerOnVM_Task, 'duration_secs': 0.585469} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.392165] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-73ab93ca-705f-42a2-b4ef-0196562a1ed6 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 976.392376] env[62569]: DEBUG nova.compute.manager [None req-73ab93ca-705f-42a2-b4ef-0196562a1ed6 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 976.393178] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8d1ee0c-9b0e-4971-8d15-3f314e368254 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.424914] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "b6e5eefc-8c06-445b-a3af-9404578b6179" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.425269] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "b6e5eefc-8c06-445b-a3af-9404578b6179" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.425555] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "b6e5eefc-8c06-445b-a3af-9404578b6179-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.425848] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "b6e5eefc-8c06-445b-a3af-9404578b6179-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.426111] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "b6e5eefc-8c06-445b-a3af-9404578b6179-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.429373] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69579c27-1cbd-4d4b-bfa6-24b302a7c3b9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.432230] env[62569]: INFO nova.compute.manager [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Terminating instance [ 976.439747] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6924b93-0392-44cb-accd-2eb0313fde77 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.470818] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fedebe9-869f-4fb5-be7c-224defea005b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.478263] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bad7beb6-a846-4a14-a206-23311f3d5d54 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.491446] env[62569]: DEBUG nova.compute.provider_tree [None req-262450e9-79ae-4558-88f2-16108de5835a tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.499234] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "refresh_cache-d56faf39-6caf-4489-98d7-342fc5fb7d40" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.499377] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquired lock "refresh_cache-d56faf39-6caf-4489-98d7-342fc5fb7d40" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.499525] env[62569]: DEBUG nova.network.neutron [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 976.510030] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250500, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.534804] env[62569]: DEBUG oslo_concurrency.lockutils [req-ab9fedfc-1146-4c8d-af46-04a0e11b9235 req-6ee82fab-535d-4627-be1b-49249c08169f service nova] Releasing lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.535242] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]522fe663-30b1-6c64-b93b-1bb78c4f7899, 'name': SearchDatastore_Task, 'duration_secs': 0.014637} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.535472] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.535659] env[62569]: DEBUG nova.network.neutron [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 976.537794] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ac29c79-8670-4d93-97bf-cf13da9fd709 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.543771] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 976.543771] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527d2975-9a58-2b08-d547-cdc2c383575a" [ 976.543771] env[62569]: _type = "Task" [ 976.543771] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.552454] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527d2975-9a58-2b08-d547-cdc2c383575a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.940085] env[62569]: DEBUG nova.compute.manager [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 976.940363] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 976.941269] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d25a5c88-2be9-4e63-9597-270bab3ef4ba {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.949133] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 976.949416] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2c862bc8-5273-42af-adeb-d5b77e95416f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.956159] env[62569]: DEBUG oslo_vmware.api [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 976.956159] env[62569]: value = "task-1250501" [ 976.956159] env[62569]: _type = "Task" [ 976.956159] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.966497] env[62569]: DEBUG oslo_vmware.api [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250501, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.995996] env[62569]: DEBUG nova.scheduler.client.report [None req-262450e9-79ae-4558-88f2-16108de5835a tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 977.015047] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250500, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.933681} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.018257] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 35350164-e295-40ea-85a6-e8e79c630e0f/35350164-e295-40ea-85a6-e8e79c630e0f.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 977.018577] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 977.019055] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bf6333ba-f1d0-43d0-b6b5-6abc945ff6b6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.025914] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 977.025914] env[62569]: value = "task-1250502" [ 977.025914] env[62569]: _type = "Task" [ 977.025914] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.034613] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250502, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.044424] env[62569]: DEBUG nova.network.neutron [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 977.056566] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527d2975-9a58-2b08-d547-cdc2c383575a, 'name': SearchDatastore_Task, 'duration_secs': 0.025586} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.056696] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.057410] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] bcc7f73b-083b-477e-bceb-4f984a95d219/bcc7f73b-083b-477e-bceb-4f984a95d219.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 977.057623] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0379cc04-fe18-4b01-af17-e661fcc8ca94 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.064434] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 977.064434] env[62569]: value = "task-1250503" [ 977.064434] env[62569]: _type = "Task" [ 977.064434] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.071596] env[62569]: WARNING nova.network.neutron [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] 29fcbfe8-35ce-4e8d-874b-2b14ba720adf already exists in list: networks containing: ['29fcbfe8-35ce-4e8d-874b-2b14ba720adf']. ignoring it [ 977.071902] env[62569]: WARNING nova.network.neutron [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] 29fcbfe8-35ce-4e8d-874b-2b14ba720adf already exists in list: networks containing: ['29fcbfe8-35ce-4e8d-874b-2b14ba720adf']. ignoring it [ 977.072135] env[62569]: WARNING nova.network.neutron [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] 0d27f2d0-6f98-4031-97c8-33df605f1bb0 already exists in list: port_ids containing: ['0d27f2d0-6f98-4031-97c8-33df605f1bb0']. ignoring it [ 977.077276] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250503, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.206973] env[62569]: DEBUG nova.network.neutron [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Updating instance_info_cache with network_info: [{"id": "b6625a0e-7282-4ade-b7f0-3e5374ec8e64", "address": "fa:16:3e:cd:26:a6", "network": {"id": "334a2e22-0249-48e3-a2f5-ac37df9771d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1048653043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f7fb3e698324f8e851aefbee74ac669", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f880ac2e-d532-4f54-87bb-998a8d1bca78", "external-id": "nsx-vlan-transportzone-491", "segmentation_id": 491, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6625a0e-72", "ovs_interfaceid": "b6625a0e-7282-4ade-b7f0-3e5374ec8e64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.367926] env[62569]: DEBUG nova.objects.instance [None req-c4a8626b-2969-42e9-b440-ebbfc5417e5d tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lazy-loading 'flavor' on Instance uuid 98089a34-074b-4bdb-92ae-f9e23d2551f1 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 977.469513] env[62569]: DEBUG oslo_vmware.api [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250501, 'name': PowerOffVM_Task, 'duration_secs': 0.245915} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.469897] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 977.470181] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 977.470618] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bfbd6604-c5de-4d4d-984d-9822ee170f82 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.495530] env[62569]: DEBUG nova.network.neutron [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Updating instance_info_cache with network_info: [{"id": "6ca4b28a-1255-4f88-a7b6-54b0378768b1", "address": "fa:16:3e:03:e7:ad", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ca4b28a-12", "ovs_interfaceid": "6ca4b28a-1255-4f88-a7b6-54b0378768b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0d27f2d0-6f98-4031-97c8-33df605f1bb0", "address": "fa:16:3e:e3:fb:62", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d27f2d0-6f", "ovs_interfaceid": "0d27f2d0-6f98-4031-97c8-33df605f1bb0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.539644] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250502, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.09773} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.540303] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 977.541258] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85c447e6-4ec2-48a6-b19e-03d6da66b2b1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.570057] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 35350164-e295-40ea-85a6-e8e79c630e0f/35350164-e295-40ea-85a6-e8e79c630e0f.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 977.572754] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-72257631-6bcd-4060-ba47-f619075891f1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.587442] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 977.587661] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 977.587842] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Deleting the datastore file [datastore1] b6e5eefc-8c06-445b-a3af-9404578b6179 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 977.594651] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8028c1da-f8da-4b65-b67f-a4d9e379d4b5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.596540] env[62569]: DEBUG oslo_concurrency.lockutils [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "2ce7e7ab-8a07-468f-9f9d-643d814981bd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.596980] env[62569]: DEBUG oslo_concurrency.lockutils [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "2ce7e7ab-8a07-468f-9f9d-643d814981bd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.605159] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250503, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.607780] env[62569]: DEBUG oslo_vmware.api [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for the task: (returnval){ [ 977.607780] env[62569]: value = "task-1250506" [ 977.607780] env[62569]: _type = "Task" [ 977.607780] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.608114] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 977.608114] env[62569]: value = "task-1250505" [ 977.608114] env[62569]: _type = "Task" [ 977.608114] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.621975] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250505, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.625778] env[62569]: DEBUG oslo_vmware.api [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250506, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.709427] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Releasing lock "refresh_cache-d56faf39-6caf-4489-98d7-342fc5fb7d40" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.709860] env[62569]: DEBUG nova.compute.manager [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Instance network_info: |[{"id": "b6625a0e-7282-4ade-b7f0-3e5374ec8e64", "address": "fa:16:3e:cd:26:a6", "network": {"id": "334a2e22-0249-48e3-a2f5-ac37df9771d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1048653043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f7fb3e698324f8e851aefbee74ac669", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f880ac2e-d532-4f54-87bb-998a8d1bca78", "external-id": "nsx-vlan-transportzone-491", "segmentation_id": 491, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6625a0e-72", "ovs_interfaceid": "b6625a0e-7282-4ade-b7f0-3e5374ec8e64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 977.710367] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cd:26:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f880ac2e-d532-4f54-87bb-998a8d1bca78', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b6625a0e-7282-4ade-b7f0-3e5374ec8e64', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 977.719653] env[62569]: DEBUG oslo.service.loopingcall [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 977.719939] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 977.720218] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-35db3d29-a87b-496a-b671-611bfea01423 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.743623] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 977.743623] env[62569]: value = "task-1250507" [ 977.743623] env[62569]: _type = "Task" [ 977.743623] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.757931] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250507, 'name': CreateVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.873181] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c4a8626b-2969-42e9-b440-ebbfc5417e5d tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.339s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.998930] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.999966] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.000249] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.000682] env[62569]: DEBUG oslo_concurrency.lockutils [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] Acquired lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.000937] env[62569]: DEBUG nova.network.neutron [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Refreshing network info cache for port 6ca4b28a-1255-4f88-a7b6-54b0378768b1 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 978.002797] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0384e3d5-4fac-4879-909f-b710e3b0633a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.007474] env[62569]: DEBUG oslo_concurrency.lockutils [None req-262450e9-79ae-4558-88f2-16108de5835a tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.253s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.024916] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.941s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.026687] env[62569]: INFO nova.compute.claims [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 978.032933] env[62569]: DEBUG nova.virt.hardware [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 978.033212] env[62569]: DEBUG nova.virt.hardware [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 978.033375] env[62569]: DEBUG nova.virt.hardware [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 978.033707] env[62569]: DEBUG nova.virt.hardware [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 978.033935] env[62569]: DEBUG nova.virt.hardware [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 978.034130] env[62569]: DEBUG nova.virt.hardware [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 978.034445] env[62569]: DEBUG nova.virt.hardware [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 978.034940] env[62569]: DEBUG nova.virt.hardware [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 978.035217] env[62569]: DEBUG nova.virt.hardware [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 978.035448] env[62569]: DEBUG nova.virt.hardware [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 978.035685] env[62569]: DEBUG nova.virt.hardware [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 978.043269] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Reconfiguring VM to attach interface {{(pid=62569) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 978.044838] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4738e895-8eb7-45e6-bc69-e4dae088f04f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.066828] env[62569]: DEBUG oslo_vmware.api [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 978.066828] env[62569]: value = "task-1250508" [ 978.066828] env[62569]: _type = "Task" [ 978.066828] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.072967] env[62569]: DEBUG nova.compute.manager [req-f062461a-c438-46ac-8758-49bf30ac1f1f req-35512c9b-2abd-413f-843b-145d1f8b93ae service nova] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Received event network-changed-b6625a0e-7282-4ade-b7f0-3e5374ec8e64 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 978.073231] env[62569]: DEBUG nova.compute.manager [req-f062461a-c438-46ac-8758-49bf30ac1f1f req-35512c9b-2abd-413f-843b-145d1f8b93ae service nova] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Refreshing instance network info cache due to event network-changed-b6625a0e-7282-4ade-b7f0-3e5374ec8e64. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 978.073490] env[62569]: DEBUG oslo_concurrency.lockutils [req-f062461a-c438-46ac-8758-49bf30ac1f1f req-35512c9b-2abd-413f-843b-145d1f8b93ae service nova] Acquiring lock "refresh_cache-d56faf39-6caf-4489-98d7-342fc5fb7d40" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.073717] env[62569]: DEBUG oslo_concurrency.lockutils [req-f062461a-c438-46ac-8758-49bf30ac1f1f req-35512c9b-2abd-413f-843b-145d1f8b93ae service nova] Acquired lock "refresh_cache-d56faf39-6caf-4489-98d7-342fc5fb7d40" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.073908] env[62569]: DEBUG nova.network.neutron [req-f062461a-c438-46ac-8758-49bf30ac1f1f req-35512c9b-2abd-413f-843b-145d1f8b93ae service nova] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Refreshing network info cache for port b6625a0e-7282-4ade-b7f0-3e5374ec8e64 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 978.085026] env[62569]: DEBUG oslo_vmware.api [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250508, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.090623] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250503, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.747359} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.090912] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] bcc7f73b-083b-477e-bceb-4f984a95d219/bcc7f73b-083b-477e-bceb-4f984a95d219.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 978.091182] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 978.091708] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b8c45b48-4ebd-4d83-830d-d9ef0b1ce189 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.098861] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 978.098861] env[62569]: value = "task-1250509" [ 978.098861] env[62569]: _type = "Task" [ 978.098861] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.102798] env[62569]: DEBUG nova.compute.manager [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 978.111037] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250509, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.122877] env[62569]: DEBUG oslo_vmware.api [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Task: {'id': task-1250506, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.421945} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.126309] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 978.126309] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 978.128189] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 978.128189] env[62569]: INFO nova.compute.manager [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Took 1.19 seconds to destroy the instance on the hypervisor. [ 978.128189] env[62569]: DEBUG oslo.service.loopingcall [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 978.128189] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250505, 'name': ReconfigVM_Task, 'duration_secs': 0.367157} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.128189] env[62569]: DEBUG nova.compute.manager [-] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 978.128189] env[62569]: DEBUG nova.network.neutron [-] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 978.128965] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 35350164-e295-40ea-85a6-e8e79c630e0f/35350164-e295-40ea-85a6-e8e79c630e0f.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 978.129719] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5e4d1bc6-a61b-4bd9-8967-491d5302a7b0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.136123] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 978.136123] env[62569]: value = "task-1250510" [ 978.136123] env[62569]: _type = "Task" [ 978.136123] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.147569] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250510, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.253247] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250507, 'name': CreateVM_Task, 'duration_secs': 0.365196} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.253436] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 978.254158] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.254336] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.254663] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 978.254926] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f886c13-5495-4926-b155-961ef2b606dc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.262103] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 978.262103] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52febf93-2d98-4791-e531-ae83582577b9" [ 978.262103] env[62569]: _type = "Task" [ 978.262103] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.273074] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52febf93-2d98-4791-e531-ae83582577b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.290388] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.290540] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquired lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.290699] env[62569]: DEBUG nova.network.neutron [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Forcefully refreshing network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 978.405600] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "b47cd2d7-0cd2-41af-8ed1-a6dfca323516" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.405873] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "b47cd2d7-0cd2-41af-8ed1-a6dfca323516" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.406110] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "b47cd2d7-0cd2-41af-8ed1-a6dfca323516-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.406308] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "b47cd2d7-0cd2-41af-8ed1-a6dfca323516-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.406500] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "b47cd2d7-0cd2-41af-8ed1-a6dfca323516-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.408944] env[62569]: INFO nova.compute.manager [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Terminating instance [ 978.583808] env[62569]: DEBUG oslo_vmware.api [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250508, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.609248] env[62569]: INFO nova.scheduler.client.report [None req-262450e9-79ae-4558-88f2-16108de5835a tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Deleted allocation for migration 5b4b811b-c7c1-4e18-b28c-14df9b5a62ed [ 978.626870] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250509, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.110756} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.630308] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 978.633836] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a373143-152a-4042-947d-6a60a4d26301 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.662070] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] bcc7f73b-083b-477e-bceb-4f984a95d219/bcc7f73b-083b-477e-bceb-4f984a95d219.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 978.663551] env[62569]: DEBUG oslo_concurrency.lockutils [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.664906] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f8bb64e-b3be-4b9a-b93c-835dd097abbe {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.688226] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250510, 'name': Rename_Task, 'duration_secs': 0.184602} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.689492] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 978.689794] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0a61a6e8-df86-47fa-9e9f-c3f1e20d67a4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.694087] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 978.694087] env[62569]: value = "task-1250511" [ 978.694087] env[62569]: _type = "Task" [ 978.694087] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.699194] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 978.699194] env[62569]: value = "task-1250512" [ 978.699194] env[62569]: _type = "Task" [ 978.699194] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.705973] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250511, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.711367] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250512, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.774593] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52febf93-2d98-4791-e531-ae83582577b9, 'name': SearchDatastore_Task, 'duration_secs': 0.019006} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.774913] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.775308] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 978.775665] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.775932] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.776252] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 978.776641] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3bf275e4-43c5-4d7f-8b9a-d0b4bb4b46dd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.787419] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 978.787625] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 978.788448] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6fab68e5-7a7f-4545-98cc-b96136445626 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.795875] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 978.795875] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52e48e8c-ec46-66b2-f07c-895f423ea338" [ 978.795875] env[62569]: _type = "Task" [ 978.795875] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.804197] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52e48e8c-ec46-66b2-f07c-895f423ea338, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.848244] env[62569]: DEBUG nova.network.neutron [req-f062461a-c438-46ac-8758-49bf30ac1f1f req-35512c9b-2abd-413f-843b-145d1f8b93ae service nova] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Updated VIF entry in instance network info cache for port b6625a0e-7282-4ade-b7f0-3e5374ec8e64. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 978.848712] env[62569]: DEBUG nova.network.neutron [req-f062461a-c438-46ac-8758-49bf30ac1f1f req-35512c9b-2abd-413f-843b-145d1f8b93ae service nova] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Updating instance_info_cache with network_info: [{"id": "b6625a0e-7282-4ade-b7f0-3e5374ec8e64", "address": "fa:16:3e:cd:26:a6", "network": {"id": "334a2e22-0249-48e3-a2f5-ac37df9771d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1048653043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f7fb3e698324f8e851aefbee74ac669", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f880ac2e-d532-4f54-87bb-998a8d1bca78", "external-id": "nsx-vlan-transportzone-491", "segmentation_id": 491, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6625a0e-72", "ovs_interfaceid": "b6625a0e-7282-4ade-b7f0-3e5374ec8e64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.882521] env[62569]: DEBUG nova.network.neutron [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Updated VIF entry in instance network info cache for port 6ca4b28a-1255-4f88-a7b6-54b0378768b1. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 978.882982] env[62569]: DEBUG nova.network.neutron [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Updating instance_info_cache with network_info: [{"id": "6ca4b28a-1255-4f88-a7b6-54b0378768b1", "address": "fa:16:3e:03:e7:ad", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ca4b28a-12", "ovs_interfaceid": "6ca4b28a-1255-4f88-a7b6-54b0378768b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0d27f2d0-6f98-4031-97c8-33df605f1bb0", "address": "fa:16:3e:e3:fb:62", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0d27f2d0-6f", "ovs_interfaceid": "0d27f2d0-6f98-4031-97c8-33df605f1bb0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.913033] env[62569]: DEBUG nova.compute.manager [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 978.913352] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 978.914406] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e5b125-ae80-477b-8536-52b4841d9c29 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.923257] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 978.923516] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f7a42e07-5498-46b6-aeac-f1320ede1021 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.929011] env[62569]: DEBUG oslo_vmware.api [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 978.929011] env[62569]: value = "task-1250513" [ 978.929011] env[62569]: _type = "Task" [ 978.929011] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.936418] env[62569]: DEBUG oslo_vmware.api [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250513, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.009584] env[62569]: DEBUG nova.network.neutron [-] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.079699] env[62569]: DEBUG oslo_vmware.api [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250508, 'name': ReconfigVM_Task, 'duration_secs': 0.57095} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.082614] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.082882] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Reconfigured VM to attach interface {{(pid=62569) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 979.122708] env[62569]: DEBUG oslo_concurrency.lockutils [None req-262450e9-79ae-4558-88f2-16108de5835a tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "09eeb937-aabc-418d-9fd3-3c7a72f4ef75" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 12.900s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.207847] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250511, 'name': ReconfigVM_Task, 'duration_secs': 0.427725} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.211577] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Reconfigured VM instance instance-00000060 to attach disk [datastore2] bcc7f73b-083b-477e-bceb-4f984a95d219/bcc7f73b-083b-477e-bceb-4f984a95d219.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 979.212754] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c831755e-59b0-4746-9381-661f21cc6ed9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.217391] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250512, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.222808] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 979.222808] env[62569]: value = "task-1250514" [ 979.222808] env[62569]: _type = "Task" [ 979.222808] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.234636] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250514, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.249043] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-455fd653-f22e-4a4d-b7c7-0c9ecfc522e0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.256719] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e047eeb-0fab-4cc6-b7a1-427b4ba189cf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.287586] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f436c321-773c-4aec-ae54-ecc2034a8370 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.295635] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7a36a6f-2285-4f93-a83b-18df3671125b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.312890] env[62569]: DEBUG nova.compute.provider_tree [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 979.318110] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca86470d-1e63-402f-bc09-a2adddc7d62a tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.318300] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca86470d-1e63-402f-bc09-a2adddc7d62a tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.319717] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52e48e8c-ec46-66b2-f07c-895f423ea338, 'name': SearchDatastore_Task, 'duration_secs': 0.014831} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.321202] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ffe0d86-5283-4205-8012-44fd1cf7580d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.329338] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 979.329338] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520f22ce-0919-e4c1-7a37-789dea08b8c5" [ 979.329338] env[62569]: _type = "Task" [ 979.329338] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.338031] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520f22ce-0919-e4c1-7a37-789dea08b8c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.351723] env[62569]: DEBUG oslo_concurrency.lockutils [req-f062461a-c438-46ac-8758-49bf30ac1f1f req-35512c9b-2abd-413f-843b-145d1f8b93ae service nova] Releasing lock "refresh_cache-d56faf39-6caf-4489-98d7-342fc5fb7d40" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.385581] env[62569]: DEBUG oslo_concurrency.lockutils [req-7b4b9aec-d532-4db0-95b2-0c10df091e66 req-4be0c3ba-84ff-4100-a3af-e8dafb564b6c service nova] Releasing lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.440661] env[62569]: DEBUG oslo_vmware.api [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250513, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.512405] env[62569]: INFO nova.compute.manager [-] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Took 1.38 seconds to deallocate network for instance. [ 979.563936] env[62569]: DEBUG nova.network.neutron [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Updating instance_info_cache with network_info: [{"id": "90b3aaef-2e72-486f-be8d-a4202a6cb0c6", "address": "fa:16:3e:0c:8f:dd", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90b3aaef-2e", "ovs_interfaceid": "90b3aaef-2e72-486f-be8d-a4202a6cb0c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.587246] env[62569]: DEBUG oslo_concurrency.lockutils [None req-d50832db-cfe1-4c6c-ae2c-858d04457053 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "interface-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e-0d27f2d0-6f98-4031-97c8-33df605f1bb0" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.399s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.631180] env[62569]: DEBUG oslo_concurrency.lockutils [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "09eeb937-aabc-418d-9fd3-3c7a72f4ef75" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.631462] env[62569]: DEBUG oslo_concurrency.lockutils [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "09eeb937-aabc-418d-9fd3-3c7a72f4ef75" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.631679] env[62569]: DEBUG oslo_concurrency.lockutils [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "09eeb937-aabc-418d-9fd3-3c7a72f4ef75-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.631869] env[62569]: DEBUG oslo_concurrency.lockutils [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "09eeb937-aabc-418d-9fd3-3c7a72f4ef75-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.632056] env[62569]: DEBUG oslo_concurrency.lockutils [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "09eeb937-aabc-418d-9fd3-3c7a72f4ef75-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.634317] env[62569]: INFO nova.compute.manager [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Terminating instance [ 979.710385] env[62569]: DEBUG oslo_vmware.api [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250512, 'name': PowerOnVM_Task, 'duration_secs': 0.548234} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.710666] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 979.710875] env[62569]: INFO nova.compute.manager [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Took 9.33 seconds to spawn the instance on the hypervisor. [ 979.711073] env[62569]: DEBUG nova.compute.manager [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 979.711899] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f2af735-4883-4c91-9b50-ff0a2ca05bbf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.733249] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250514, 'name': Rename_Task, 'duration_secs': 0.203241} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.733510] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 979.733766] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-62635089-b9cf-43fe-88bb-0e826382e837 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.740399] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 979.740399] env[62569]: value = "task-1250515" [ 979.740399] env[62569]: _type = "Task" [ 979.740399] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.747629] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250515, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.824740] env[62569]: DEBUG nova.compute.utils [None req-ca86470d-1e63-402f-bc09-a2adddc7d62a tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 979.839501] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520f22ce-0919-e4c1-7a37-789dea08b8c5, 'name': SearchDatastore_Task, 'duration_secs': 0.01799} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.839792] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.840080] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] d56faf39-6caf-4489-98d7-342fc5fb7d40/d56faf39-6caf-4489-98d7-342fc5fb7d40.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 979.840319] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-350efbbe-c7eb-47cb-afbb-b44ca42b2f0e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.845726] env[62569]: ERROR nova.scheduler.client.report [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [req-b13ee4b2-dffb-433d-8112-75bcfbdf69b7] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID fa06556a-5785-4014-b8bd-bc240a0cf716. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b13ee4b2-dffb-433d-8112-75bcfbdf69b7"}]} [ 979.849568] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 979.849568] env[62569]: value = "task-1250516" [ 979.849568] env[62569]: _type = "Task" [ 979.849568] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.857127] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250516, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.861947] env[62569]: DEBUG nova.scheduler.client.report [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Refreshing inventories for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 979.877493] env[62569]: DEBUG nova.scheduler.client.report [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Updating ProviderTree inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 979.877730] env[62569]: DEBUG nova.compute.provider_tree [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 979.890487] env[62569]: DEBUG nova.scheduler.client.report [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Refreshing aggregate associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, aggregates: None {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 979.908871] env[62569]: DEBUG nova.scheduler.client.report [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Refreshing trait associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 979.939425] env[62569]: DEBUG oslo_vmware.api [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250513, 'name': PowerOffVM_Task, 'duration_secs': 0.604157} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.939711] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 979.939906] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 979.940178] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1d4688f2-4c28-46dc-a3b9-f619c5ed939a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.008543] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 980.008806] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 980.008994] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Deleting the datastore file [datastore1] b47cd2d7-0cd2-41af-8ed1-a6dfca323516 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 980.009345] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-198036ab-3a08-4ce1-9fa9-dacd7f40631c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.019788] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.022034] env[62569]: DEBUG oslo_vmware.api [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 980.022034] env[62569]: value = "task-1250518" [ 980.022034] env[62569]: _type = "Task" [ 980.022034] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.036544] env[62569]: DEBUG oslo_vmware.api [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250518, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.066142] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Releasing lock "refresh_cache-b47cd2d7-0cd2-41af-8ed1-a6dfca323516" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.066376] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Updated the network info_cache for instance {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10379}} [ 980.066621] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 980.066808] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 980.066958] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 980.067123] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 980.067269] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 980.067418] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 980.067548] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62569) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 980.067690] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 980.100507] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f04606-2edd-482e-be42-6a34b9eff3ee {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.110325] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e66dd068-d643-4019-8b6c-4a455fda8a45 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.147498] env[62569]: DEBUG nova.compute.manager [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 980.147739] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 980.148916] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f8a8c7c-370d-48d5-b4a1-8c5f92c19542 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.152494] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab62403-83eb-456b-9561-ab0b5012ef95 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.161378] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a07abe6b-f2bc-44ea-b293-4e77fd8df1de {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.168037] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 980.168793] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f4979a1-229c-41b0-bdf5-feafb6e2b45a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.179192] env[62569]: DEBUG nova.compute.provider_tree [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 980.187981] env[62569]: DEBUG oslo_vmware.api [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 980.187981] env[62569]: value = "task-1250519" [ 980.187981] env[62569]: _type = "Task" [ 980.187981] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.201265] env[62569]: DEBUG oslo_vmware.api [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250519, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.229087] env[62569]: INFO nova.compute.manager [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Took 14.25 seconds to build instance. [ 980.253094] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250515, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.328304] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca86470d-1e63-402f-bc09-a2adddc7d62a tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.359765] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250516, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.440506} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.360391] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] d56faf39-6caf-4489-98d7-342fc5fb7d40/d56faf39-6caf-4489-98d7-342fc5fb7d40.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 980.360391] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 980.360557] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5dc53672-f3d1-409f-be6e-f1f05f22e7ca {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.370891] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 980.370891] env[62569]: value = "task-1250520" [ 980.370891] env[62569]: _type = "Task" [ 980.370891] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.381647] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250520, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.531637] env[62569]: DEBUG oslo_vmware.api [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250518, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.303366} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.531920] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 980.532129] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 980.532317] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 980.532511] env[62569]: INFO nova.compute.manager [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Took 1.62 seconds to destroy the instance on the hypervisor. [ 980.532765] env[62569]: DEBUG oslo.service.loopingcall [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 980.532960] env[62569]: DEBUG nova.compute.manager [-] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 980.533068] env[62569]: DEBUG nova.network.neutron [-] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 980.570834] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.700935] env[62569]: DEBUG oslo_vmware.api [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250519, 'name': PowerOffVM_Task, 'duration_secs': 0.254466} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.701832] env[62569]: ERROR nova.scheduler.client.report [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [req-7bde98a5-517d-4339-ad40-ce13255f2ebb] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID fa06556a-5785-4014-b8bd-bc240a0cf716. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-7bde98a5-517d-4339-ad40-ce13255f2ebb"}]} [ 980.702300] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 980.702498] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 980.704597] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3e6869db-37ea-4810-a5eb-3bee811255d8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.720560] env[62569]: DEBUG nova.scheduler.client.report [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Refreshing inventories for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 980.731043] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1791501f-5b70-4cca-8817-8a97e7ef6055 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "35350164-e295-40ea-85a6-e8e79c630e0f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.760s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.735095] env[62569]: DEBUG nova.scheduler.client.report [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Updating ProviderTree inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 980.735324] env[62569]: DEBUG nova.compute.provider_tree [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 980.745995] env[62569]: DEBUG nova.scheduler.client.report [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Refreshing aggregate associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, aggregates: None {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 980.753077] env[62569]: DEBUG oslo_vmware.api [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250515, 'name': PowerOnVM_Task, 'duration_secs': 0.650365} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.753339] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 980.753545] env[62569]: DEBUG nova.compute.manager [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 980.754353] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-768afeba-f400-45ea-8f98-ab3ee0eadfdf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.768556] env[62569]: DEBUG nova.scheduler.client.report [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Refreshing trait associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 980.881958] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250520, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067974} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.884752] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 980.885849] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48c01804-86d8-42e7-bcb2-33818f4fbd1d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.893163] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 980.893402] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 980.893605] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Deleting the datastore file [datastore2] 09eeb937-aabc-418d-9fd3-3c7a72f4ef75 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 980.902665] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4334fb7b-ec35-4d9d-a80c-10cb800e4b31 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.913143] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] d56faf39-6caf-4489-98d7-342fc5fb7d40/d56faf39-6caf-4489-98d7-342fc5fb7d40.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 980.916928] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b33de1f0-dda8-4553-b166-35b96e077de1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.941020] env[62569]: DEBUG oslo_vmware.api [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 980.941020] env[62569]: value = "task-1250522" [ 980.941020] env[62569]: _type = "Task" [ 980.941020] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.941020] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 980.941020] env[62569]: value = "task-1250523" [ 980.941020] env[62569]: _type = "Task" [ 980.941020] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.955262] env[62569]: DEBUG oslo_vmware.api [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250522, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.958009] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250523, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.031374] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e70d36cd-552c-45bf-9a71-2d82969d9bf2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.039347] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8fad9f9-98ab-4f6e-ab60-46b9f7edb32f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.073059] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-686a60c0-eae6-4992-bfd6-7123fefc66ad {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.080526] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c557284-3ae2-4d0f-8cba-f061a89b2ac7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.094987] env[62569]: DEBUG nova.compute.provider_tree [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 981.174623] env[62569]: DEBUG oslo_concurrency.lockutils [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "4df2078e-d7ce-4a3f-904b-f3303071ffff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.174862] env[62569]: DEBUG oslo_concurrency.lockutils [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "4df2078e-d7ce-4a3f-904b-f3303071ffff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.271670] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.334937] env[62569]: DEBUG nova.compute.manager [req-0bb22301-4c10-4344-9556-c104e8d4b196 req-8a2c92c0-fe25-42e4-a99f-2c9cadc92802 service nova] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Received event network-vif-deleted-cfddbae7-9ab0-4bc6-aea1-46a0cde743fc {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 981.434750] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca86470d-1e63-402f-bc09-a2adddc7d62a tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.435052] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca86470d-1e63-402f-bc09-a2adddc7d62a tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.435368] env[62569]: INFO nova.compute.manager [None req-ca86470d-1e63-402f-bc09-a2adddc7d62a tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Attaching volume 457ddfd9-7527-4e8c-8e6e-7428d784f6d3 to /dev/sdc [ 981.454035] env[62569]: DEBUG oslo_vmware.api [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250522, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144341} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.457172] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 981.457615] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 981.457615] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 981.458020] env[62569]: INFO nova.compute.manager [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Took 1.31 seconds to destroy the instance on the hypervisor. [ 981.458020] env[62569]: DEBUG oslo.service.loopingcall [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 981.458385] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250523, 'name': ReconfigVM_Task, 'duration_secs': 0.313335} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.458467] env[62569]: DEBUG nova.compute.manager [-] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 981.458522] env[62569]: DEBUG nova.network.neutron [-] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 981.460157] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Reconfigured VM instance instance-00000062 to attach disk [datastore2] d56faf39-6caf-4489-98d7-342fc5fb7d40/d56faf39-6caf-4489-98d7-342fc5fb7d40.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 981.460835] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b226903a-d237-4d5d-8dcb-4137ab95d99f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.468105] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 981.468105] env[62569]: value = "task-1250524" [ 981.468105] env[62569]: _type = "Task" [ 981.468105] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.477269] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250524, 'name': Rename_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.485155] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b06bd84e-001b-4114-9f02-dfa2f1be9090 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.494015] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dce7a79f-8e66-48df-beaf-66ea6862651d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.507321] env[62569]: DEBUG nova.virt.block_device [None req-ca86470d-1e63-402f-bc09-a2adddc7d62a tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Updating existing volume attachment record: b9df3136-850d-4373-ab1d-202ddf1d86dc {{(pid=62569) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 981.597792] env[62569]: DEBUG nova.scheduler.client.report [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 981.677292] env[62569]: DEBUG nova.compute.manager [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 981.983631] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250524, 'name': Rename_Task, 'duration_secs': 0.147711} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.984663] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 981.985627] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2dd0757a-f0aa-47af-91ac-0efae9998827 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.991456] env[62569]: DEBUG nova.network.neutron [-] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.995471] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 981.995471] env[62569]: value = "task-1250526" [ 981.995471] env[62569]: _type = "Task" [ 981.995471] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.008422] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7e08333f-03b3-43e7-bf31-bbb13fb4796b tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "35350164-e295-40ea-85a6-e8e79c630e0f" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.008866] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7e08333f-03b3-43e7-bf31-bbb13fb4796b tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "35350164-e295-40ea-85a6-e8e79c630e0f" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.009190] env[62569]: DEBUG nova.compute.manager [None req-7e08333f-03b3-43e7-bf31-bbb13fb4796b tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 982.010091] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250526, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.011686] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29512112-9eeb-4276-bf1f-1cf2639dfe64 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.019047] env[62569]: DEBUG nova.compute.manager [None req-7e08333f-03b3-43e7-bf31-bbb13fb4796b tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62569) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 982.019684] env[62569]: DEBUG nova.objects.instance [None req-7e08333f-03b3-43e7-bf31-bbb13fb4796b tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lazy-loading 'flavor' on Instance uuid 35350164-e295-40ea-85a6-e8e79c630e0f {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.038316] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "interface-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e-0d27f2d0-6f98-4031-97c8-33df605f1bb0" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.038634] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "interface-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e-0d27f2d0-6f98-4031-97c8-33df605f1bb0" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.103885] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.079s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.104481] env[62569]: DEBUG nova.compute.manager [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 982.107168] env[62569]: DEBUG oslo_concurrency.lockutils [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.444s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.108800] env[62569]: INFO nova.compute.claims [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 982.202078] env[62569]: DEBUG oslo_concurrency.lockutils [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.418432] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "bcc7f73b-083b-477e-bceb-4f984a95d219" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.418713] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "bcc7f73b-083b-477e-bceb-4f984a95d219" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.418930] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "bcc7f73b-083b-477e-bceb-4f984a95d219-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.419139] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "bcc7f73b-083b-477e-bceb-4f984a95d219-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.419328] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "bcc7f73b-083b-477e-bceb-4f984a95d219-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.421595] env[62569]: INFO nova.compute.manager [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Terminating instance [ 982.428627] env[62569]: DEBUG nova.network.neutron [-] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.497573] env[62569]: INFO nova.compute.manager [-] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Took 1.96 seconds to deallocate network for instance. [ 982.512092] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250526, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.541376] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.541631] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.542693] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc5af23-3f17-4fee-aa8f-8c685d9315c2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.562102] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-330ef8f2-23cf-4428-8b4d-c9e348a88598 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.586975] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Reconfiguring VM to detach interface {{(pid=62569) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 982.587236] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e094aadb-b876-4ff8-bed4-a347fc63926a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.603901] env[62569]: DEBUG oslo_vmware.api [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 982.603901] env[62569]: value = "task-1250527" [ 982.603901] env[62569]: _type = "Task" [ 982.603901] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.611131] env[62569]: DEBUG oslo_vmware.api [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250527, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.612340] env[62569]: DEBUG nova.compute.utils [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 982.615310] env[62569]: DEBUG nova.compute.manager [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 982.615479] env[62569]: DEBUG nova.network.neutron [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 982.678319] env[62569]: DEBUG nova.policy [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a14b0267fd514378ba84cad3d7d39948', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '44464561dec24f209ec4493ac856679b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 982.925794] env[62569]: DEBUG nova.compute.manager [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 982.926106] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 982.927497] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dddfb86d-1d57-4828-814e-b8da66d8bdb1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.930738] env[62569]: INFO nova.compute.manager [-] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Took 1.47 seconds to deallocate network for instance. [ 982.937979] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 982.938572] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2e5b8dad-b0e9-4dae-9543-a39405298653 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.946341] env[62569]: DEBUG oslo_vmware.api [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 982.946341] env[62569]: value = "task-1250528" [ 982.946341] env[62569]: _type = "Task" [ 982.946341] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.952982] env[62569]: DEBUG oslo_vmware.api [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250528, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.980727] env[62569]: DEBUG nova.network.neutron [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Successfully created port: 2e9fac89-877b-4c36-808d-21c6f01c992d {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 983.013021] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.013021] env[62569]: DEBUG oslo_vmware.api [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250526, 'name': PowerOnVM_Task, 'duration_secs': 0.519388} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.013021] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 983.013021] env[62569]: INFO nova.compute.manager [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Took 7.96 seconds to spawn the instance on the hypervisor. [ 983.013021] env[62569]: DEBUG nova.compute.manager [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 983.013021] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60dbc025-d1c6-4a4f-b732-66fa4defb68a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.029698] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e08333f-03b3-43e7-bf31-bbb13fb4796b tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 983.030217] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3c78a8e5-9758-4877-981c-ae0561cb42d8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.037304] env[62569]: DEBUG oslo_vmware.api [None req-7e08333f-03b3-43e7-bf31-bbb13fb4796b tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 983.037304] env[62569]: value = "task-1250529" [ 983.037304] env[62569]: _type = "Task" [ 983.037304] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.046311] env[62569]: DEBUG oslo_vmware.api [None req-7e08333f-03b3-43e7-bf31-bbb13fb4796b tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250529, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.122619] env[62569]: DEBUG nova.compute.manager [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 983.132509] env[62569]: DEBUG oslo_vmware.api [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250527, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.347094] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1b8be92-095c-4be3-bb57-dbf077d73708 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.354759] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8863b5c1-55c7-464b-8113-d87e39cbe747 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.361723] env[62569]: DEBUG nova.compute.manager [req-8dfa2267-c540-42c6-9806-877346df2139 req-9df29a09-1947-47d0-9649-aa576642934f service nova] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Received event network-vif-deleted-90b3aaef-2e72-486f-be8d-a4202a6cb0c6 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 983.361958] env[62569]: DEBUG nova.compute.manager [req-8dfa2267-c540-42c6-9806-877346df2139 req-9df29a09-1947-47d0-9649-aa576642934f service nova] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Received event network-vif-deleted-c967324f-ab14-480b-a8b8-710db5827863 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 983.391190] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87c8aba7-4fa4-4b11-8281-632ec2f229dd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.399121] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45e7537a-d519-44f3-9240-f6d0868e56b9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.413454] env[62569]: DEBUG nova.compute.provider_tree [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 983.440622] env[62569]: DEBUG oslo_concurrency.lockutils [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.454564] env[62569]: DEBUG oslo_vmware.api [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250528, 'name': PowerOffVM_Task, 'duration_secs': 0.351449} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.454873] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 983.455096] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 983.455383] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c6a131fa-26f8-4b5b-89d2-7fdbdcdb16b6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.521825] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 983.522075] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 983.522286] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Deleting the datastore file [datastore2] bcc7f73b-083b-477e-bceb-4f984a95d219 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 983.522620] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-54858292-e0b1-4604-bce7-0e1761b48903 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.533040] env[62569]: DEBUG oslo_vmware.api [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 983.533040] env[62569]: value = "task-1250531" [ 983.533040] env[62569]: _type = "Task" [ 983.533040] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.536415] env[62569]: INFO nova.compute.manager [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Took 14.99 seconds to build instance. [ 983.544816] env[62569]: DEBUG oslo_vmware.api [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250531, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.549930] env[62569]: DEBUG oslo_vmware.api [None req-7e08333f-03b3-43e7-bf31-bbb13fb4796b tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250529, 'name': PowerOffVM_Task, 'duration_secs': 0.26402} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.550213] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e08333f-03b3-43e7-bf31-bbb13fb4796b tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 983.550407] env[62569]: DEBUG nova.compute.manager [None req-7e08333f-03b3-43e7-bf31-bbb13fb4796b tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 983.551189] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3055717-34bb-4936-8975-d4fdfd5ecbfa {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.614275] env[62569]: DEBUG oslo_vmware.api [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250527, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.917619] env[62569]: DEBUG nova.scheduler.client.report [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 984.039227] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5808a1f4-927e-43b9-8f84-50f0d3d3d05b tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "d56faf39-6caf-4489-98d7-342fc5fb7d40" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.502s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.042875] env[62569]: DEBUG oslo_vmware.api [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250531, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131871} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.043150] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 984.043347] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 984.043528] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 984.043707] env[62569]: INFO nova.compute.manager [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Took 1.12 seconds to destroy the instance on the hypervisor. [ 984.043986] env[62569]: DEBUG oslo.service.loopingcall [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 984.044203] env[62569]: DEBUG nova.compute.manager [-] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 984.044301] env[62569]: DEBUG nova.network.neutron [-] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 984.065235] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7e08333f-03b3-43e7-bf31-bbb13fb4796b tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "35350164-e295-40ea-85a6-e8e79c630e0f" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.056s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.115763] env[62569]: DEBUG oslo_vmware.api [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250527, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.139434] env[62569]: DEBUG nova.compute.manager [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 984.164414] env[62569]: DEBUG nova.virt.hardware [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 984.164697] env[62569]: DEBUG nova.virt.hardware [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 984.164864] env[62569]: DEBUG nova.virt.hardware [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 984.165069] env[62569]: DEBUG nova.virt.hardware [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 984.165233] env[62569]: DEBUG nova.virt.hardware [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 984.165388] env[62569]: DEBUG nova.virt.hardware [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 984.165602] env[62569]: DEBUG nova.virt.hardware [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 984.165818] env[62569]: DEBUG nova.virt.hardware [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 984.166029] env[62569]: DEBUG nova.virt.hardware [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 984.166208] env[62569]: DEBUG nova.virt.hardware [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 984.166393] env[62569]: DEBUG nova.virt.hardware [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 984.167276] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfe50ed4-1d28-40c3-adfd-7da6af542205 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.175418] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5019035-243f-4dbb-808e-7295f1fafba9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.422838] env[62569]: DEBUG oslo_concurrency.lockutils [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.315s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.423637] env[62569]: DEBUG nova.compute.manager [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 984.427845] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.408s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.428200] env[62569]: DEBUG nova.objects.instance [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lazy-loading 'resources' on Instance uuid b6e5eefc-8c06-445b-a3af-9404578b6179 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.558860] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "35350164-e295-40ea-85a6-e8e79c630e0f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.559160] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "35350164-e295-40ea-85a6-e8e79c630e0f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.559383] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "35350164-e295-40ea-85a6-e8e79c630e0f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.559603] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "35350164-e295-40ea-85a6-e8e79c630e0f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.559796] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "35350164-e295-40ea-85a6-e8e79c630e0f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.564903] env[62569]: INFO nova.compute.manager [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Terminating instance [ 984.618517] env[62569]: DEBUG oslo_vmware.api [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250527, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.851694] env[62569]: DEBUG nova.network.neutron [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Successfully updated port: 2e9fac89-877b-4c36-808d-21c6f01c992d {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 984.932338] env[62569]: DEBUG nova.compute.utils [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 984.936464] env[62569]: DEBUG nova.compute.manager [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 984.936640] env[62569]: DEBUG nova.network.neutron [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 984.946932] env[62569]: DEBUG nova.compute.manager [req-56e327e4-5acb-48a2-8e6f-379d65b96a58 req-eb7b8547-f869-4414-9512-d2a115398a53 service nova] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Received event network-changed-b6625a0e-7282-4ade-b7f0-3e5374ec8e64 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 984.947166] env[62569]: DEBUG nova.compute.manager [req-56e327e4-5acb-48a2-8e6f-379d65b96a58 req-eb7b8547-f869-4414-9512-d2a115398a53 service nova] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Refreshing instance network info cache due to event network-changed-b6625a0e-7282-4ade-b7f0-3e5374ec8e64. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 984.947388] env[62569]: DEBUG oslo_concurrency.lockutils [req-56e327e4-5acb-48a2-8e6f-379d65b96a58 req-eb7b8547-f869-4414-9512-d2a115398a53 service nova] Acquiring lock "refresh_cache-d56faf39-6caf-4489-98d7-342fc5fb7d40" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.947538] env[62569]: DEBUG oslo_concurrency.lockutils [req-56e327e4-5acb-48a2-8e6f-379d65b96a58 req-eb7b8547-f869-4414-9512-d2a115398a53 service nova] Acquired lock "refresh_cache-d56faf39-6caf-4489-98d7-342fc5fb7d40" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.947708] env[62569]: DEBUG nova.network.neutron [req-56e327e4-5acb-48a2-8e6f-379d65b96a58 req-eb7b8547-f869-4414-9512-d2a115398a53 service nova] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Refreshing network info cache for port b6625a0e-7282-4ade-b7f0-3e5374ec8e64 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 984.977214] env[62569]: DEBUG nova.policy [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6dd052494f244fd78050c5d26f497720', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2c6a4814c234d53bbd4844cd482fbae', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 985.070733] env[62569]: DEBUG nova.compute.manager [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 985.070998] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 985.074407] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79a61cab-a7aa-4eee-9684-fab5b11170ba {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.083985] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 985.084276] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-98caec64-fe14-4907-971f-10e209d6afd5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.119291] env[62569]: DEBUG oslo_vmware.api [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250527, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.136421] env[62569]: DEBUG nova.network.neutron [-] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.138420] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9879e04-4395-4601-be2e-1cd08ab51dad {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.146052] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6662af70-6d49-4cb4-8c48-5d4aa1b52025 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.179192] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cc53d40-b61e-4e93-a1f7-9f6f0fa9c44f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.182130] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 985.182344] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 985.182529] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Deleting the datastore file [datastore2] 35350164-e295-40ea-85a6-e8e79c630e0f {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 985.183152] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b53bbe81-2fd3-454d-b8ed-6dcd6de7df1a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.191988] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d9031c1-64bc-4c0d-9cf8-ed731251faee {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.195468] env[62569]: DEBUG oslo_vmware.api [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 985.195468] env[62569]: value = "task-1250534" [ 985.195468] env[62569]: _type = "Task" [ 985.195468] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.208840] env[62569]: DEBUG nova.compute.provider_tree [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 985.216897] env[62569]: DEBUG oslo_vmware.api [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250534, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.292246] env[62569]: DEBUG nova.network.neutron [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Successfully created port: 69a7863e-3e3c-4910-be84-9e299a18442e {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 985.354667] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquiring lock "refresh_cache-9b5a1451-af46-4cd8-8e7a-f4560d8499e6" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.354830] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquired lock "refresh_cache-9b5a1451-af46-4cd8-8e7a-f4560d8499e6" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.354971] env[62569]: DEBUG nova.network.neutron [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 985.435726] env[62569]: DEBUG nova.compute.manager [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 985.443245] env[62569]: DEBUG nova.compute.manager [req-2b1a6184-5539-4836-8ca9-b9d6b746c505 req-618d6fca-7740-4ed3-bf18-b05501bd7546 service nova] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Received event network-vif-deleted-a274afa6-f296-447a-902c-0747076a59c9 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 985.443245] env[62569]: DEBUG nova.compute.manager [req-2b1a6184-5539-4836-8ca9-b9d6b746c505 req-618d6fca-7740-4ed3-bf18-b05501bd7546 service nova] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Received event network-vif-plugged-2e9fac89-877b-4c36-808d-21c6f01c992d {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 985.443321] env[62569]: DEBUG oslo_concurrency.lockutils [req-2b1a6184-5539-4836-8ca9-b9d6b746c505 req-618d6fca-7740-4ed3-bf18-b05501bd7546 service nova] Acquiring lock "9b5a1451-af46-4cd8-8e7a-f4560d8499e6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.443463] env[62569]: DEBUG oslo_concurrency.lockutils [req-2b1a6184-5539-4836-8ca9-b9d6b746c505 req-618d6fca-7740-4ed3-bf18-b05501bd7546 service nova] Lock "9b5a1451-af46-4cd8-8e7a-f4560d8499e6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.443620] env[62569]: DEBUG oslo_concurrency.lockutils [req-2b1a6184-5539-4836-8ca9-b9d6b746c505 req-618d6fca-7740-4ed3-bf18-b05501bd7546 service nova] Lock "9b5a1451-af46-4cd8-8e7a-f4560d8499e6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.443781] env[62569]: DEBUG nova.compute.manager [req-2b1a6184-5539-4836-8ca9-b9d6b746c505 req-618d6fca-7740-4ed3-bf18-b05501bd7546 service nova] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] No waiting events found dispatching network-vif-plugged-2e9fac89-877b-4c36-808d-21c6f01c992d {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 985.443946] env[62569]: WARNING nova.compute.manager [req-2b1a6184-5539-4836-8ca9-b9d6b746c505 req-618d6fca-7740-4ed3-bf18-b05501bd7546 service nova] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Received unexpected event network-vif-plugged-2e9fac89-877b-4c36-808d-21c6f01c992d for instance with vm_state building and task_state spawning. [ 985.448201] env[62569]: DEBUG nova.compute.manager [req-2b1a6184-5539-4836-8ca9-b9d6b746c505 req-618d6fca-7740-4ed3-bf18-b05501bd7546 service nova] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Received event network-changed-2e9fac89-877b-4c36-808d-21c6f01c992d {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 985.448408] env[62569]: DEBUG nova.compute.manager [req-2b1a6184-5539-4836-8ca9-b9d6b746c505 req-618d6fca-7740-4ed3-bf18-b05501bd7546 service nova] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Refreshing instance network info cache due to event network-changed-2e9fac89-877b-4c36-808d-21c6f01c992d. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 985.448585] env[62569]: DEBUG oslo_concurrency.lockutils [req-2b1a6184-5539-4836-8ca9-b9d6b746c505 req-618d6fca-7740-4ed3-bf18-b05501bd7546 service nova] Acquiring lock "refresh_cache-9b5a1451-af46-4cd8-8e7a-f4560d8499e6" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.616555] env[62569]: DEBUG oslo_vmware.api [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250527, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.642187] env[62569]: INFO nova.compute.manager [-] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Took 1.60 seconds to deallocate network for instance. [ 985.706294] env[62569]: DEBUG oslo_vmware.api [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250534, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143814} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.706574] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 985.706793] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 985.706983] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 985.707184] env[62569]: INFO nova.compute.manager [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Took 0.64 seconds to destroy the instance on the hypervisor. [ 985.707432] env[62569]: DEBUG oslo.service.loopingcall [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 985.707634] env[62569]: DEBUG nova.compute.manager [-] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 985.707732] env[62569]: DEBUG nova.network.neutron [-] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 985.712129] env[62569]: DEBUG nova.scheduler.client.report [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 985.904020] env[62569]: DEBUG nova.network.neutron [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 985.974434] env[62569]: DEBUG nova.network.neutron [req-56e327e4-5acb-48a2-8e6f-379d65b96a58 req-eb7b8547-f869-4414-9512-d2a115398a53 service nova] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Updated VIF entry in instance network info cache for port b6625a0e-7282-4ade-b7f0-3e5374ec8e64. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 985.974880] env[62569]: DEBUG nova.network.neutron [req-56e327e4-5acb-48a2-8e6f-379d65b96a58 req-eb7b8547-f869-4414-9512-d2a115398a53 service nova] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Updating instance_info_cache with network_info: [{"id": "b6625a0e-7282-4ade-b7f0-3e5374ec8e64", "address": "fa:16:3e:cd:26:a6", "network": {"id": "334a2e22-0249-48e3-a2f5-ac37df9771d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1048653043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.148", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f7fb3e698324f8e851aefbee74ac669", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f880ac2e-d532-4f54-87bb-998a8d1bca78", "external-id": "nsx-vlan-transportzone-491", "segmentation_id": 491, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6625a0e-72", "ovs_interfaceid": "b6625a0e-7282-4ade-b7f0-3e5374ec8e64", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.061547] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca86470d-1e63-402f-bc09-a2adddc7d62a tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Volume attach. Driver type: vmdk {{(pid=62569) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 986.061828] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca86470d-1e63-402f-bc09-a2adddc7d62a tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269492', 'volume_id': '457ddfd9-7527-4e8c-8e6e-7428d784f6d3', 'name': 'volume-457ddfd9-7527-4e8c-8e6e-7428d784f6d3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '98089a34-074b-4bdb-92ae-f9e23d2551f1', 'attached_at': '', 'detached_at': '', 'volume_id': '457ddfd9-7527-4e8c-8e6e-7428d784f6d3', 'serial': '457ddfd9-7527-4e8c-8e6e-7428d784f6d3'} {{(pid=62569) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 986.062865] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb3d7b0a-ea66-45de-bdbb-10bc00058d34 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.085407] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-593552e6-23ec-4031-848d-ce6528848e4d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.118124] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca86470d-1e63-402f-bc09-a2adddc7d62a tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] volume-457ddfd9-7527-4e8c-8e6e-7428d784f6d3/volume-457ddfd9-7527-4e8c-8e6e-7428d784f6d3.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 986.120233] env[62569]: DEBUG nova.network.neutron [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Updating instance_info_cache with network_info: [{"id": "2e9fac89-877b-4c36-808d-21c6f01c992d", "address": "fa:16:3e:f3:3e:ee", "network": {"id": "73711c83-88d7-490c-bc06-44e6ed11339c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1710731049-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "44464561dec24f209ec4493ac856679b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "05b1253d-2b87-4158-9ff1-dafcf829f11f", "external-id": "nsx-vlan-transportzone-55", "segmentation_id": 55, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e9fac89-87", "ovs_interfaceid": "2e9fac89-877b-4c36-808d-21c6f01c992d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.126429] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10425e6c-945a-4266-8ea2-1d2ba5605695 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.147738] env[62569]: DEBUG oslo_vmware.api [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250527, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.149088] env[62569]: DEBUG oslo_vmware.api [None req-ca86470d-1e63-402f-bc09-a2adddc7d62a tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 986.149088] env[62569]: value = "task-1250535" [ 986.149088] env[62569]: _type = "Task" [ 986.149088] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.157753] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.158039] env[62569]: DEBUG oslo_vmware.api [None req-ca86470d-1e63-402f-bc09-a2adddc7d62a tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250535, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.219229] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.791s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.221659] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 5.651s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.221838] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.222000] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62569) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 986.222384] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 4.951s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.222676] env[62569]: DEBUG nova.objects.instance [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62569) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 986.225882] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e148936-c8f1-4fbb-8a9b-07d8ffb8218b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.234455] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c22b80b2-dda4-4cf6-a626-9414f8526aea {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.248380] env[62569]: INFO nova.scheduler.client.report [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Deleted allocations for instance b6e5eefc-8c06-445b-a3af-9404578b6179 [ 986.249864] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae838813-a355-4702-854c-774326b8133d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.258518] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57a148dc-78b4-480f-9b76-a5bf7f13796a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.288878] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180055MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=62569) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 986.289021] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.457629] env[62569]: DEBUG nova.compute.manager [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 986.480267] env[62569]: DEBUG oslo_concurrency.lockutils [req-56e327e4-5acb-48a2-8e6f-379d65b96a58 req-eb7b8547-f869-4414-9512-d2a115398a53 service nova] Releasing lock "refresh_cache-d56faf39-6caf-4489-98d7-342fc5fb7d40" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.483029] env[62569]: DEBUG nova.virt.hardware [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 986.483029] env[62569]: DEBUG nova.virt.hardware [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 986.483247] env[62569]: DEBUG nova.virt.hardware [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 986.483306] env[62569]: DEBUG nova.virt.hardware [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 986.483459] env[62569]: DEBUG nova.virt.hardware [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 986.483609] env[62569]: DEBUG nova.virt.hardware [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 986.483819] env[62569]: DEBUG nova.virt.hardware [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 986.483984] env[62569]: DEBUG nova.virt.hardware [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 986.484168] env[62569]: DEBUG nova.virt.hardware [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 986.484334] env[62569]: DEBUG nova.virt.hardware [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 986.484510] env[62569]: DEBUG nova.virt.hardware [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 986.485408] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c4dd91b-c945-4d65-a674-8fd181785bcc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.493456] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9018d99-84e3-4480-83fa-1a41381680fc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.591471] env[62569]: DEBUG nova.network.neutron [-] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.623448] env[62569]: DEBUG oslo_vmware.api [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250527, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.641499] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Releasing lock "refresh_cache-9b5a1451-af46-4cd8-8e7a-f4560d8499e6" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.641915] env[62569]: DEBUG nova.compute.manager [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Instance network_info: |[{"id": "2e9fac89-877b-4c36-808d-21c6f01c992d", "address": "fa:16:3e:f3:3e:ee", "network": {"id": "73711c83-88d7-490c-bc06-44e6ed11339c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1710731049-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "44464561dec24f209ec4493ac856679b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "05b1253d-2b87-4158-9ff1-dafcf829f11f", "external-id": "nsx-vlan-transportzone-55", "segmentation_id": 55, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e9fac89-87", "ovs_interfaceid": "2e9fac89-877b-4c36-808d-21c6f01c992d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 986.642249] env[62569]: DEBUG oslo_concurrency.lockutils [req-2b1a6184-5539-4836-8ca9-b9d6b746c505 req-618d6fca-7740-4ed3-bf18-b05501bd7546 service nova] Acquired lock "refresh_cache-9b5a1451-af46-4cd8-8e7a-f4560d8499e6" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.642458] env[62569]: DEBUG nova.network.neutron [req-2b1a6184-5539-4836-8ca9-b9d6b746c505 req-618d6fca-7740-4ed3-bf18-b05501bd7546 service nova] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Refreshing network info cache for port 2e9fac89-877b-4c36-808d-21c6f01c992d {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 986.643768] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f3:3e:ee', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '05b1253d-2b87-4158-9ff1-dafcf829f11f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2e9fac89-877b-4c36-808d-21c6f01c992d', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 986.652581] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Creating folder: Project (44464561dec24f209ec4493ac856679b). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 986.654453] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-56011a37-00b4-4412-9ed1-8bef52f6bea2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.666034] env[62569]: DEBUG oslo_vmware.api [None req-ca86470d-1e63-402f-bc09-a2adddc7d62a tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250535, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.671022] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Created folder: Project (44464561dec24f209ec4493ac856679b) in parent group-v269330. [ 986.671022] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Creating folder: Instances. Parent ref: group-v269493. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 986.671022] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-77712327-386e-4bc4-8589-0730d816634b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.677853] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Created folder: Instances in parent group-v269493. [ 986.678217] env[62569]: DEBUG oslo.service.loopingcall [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 986.678366] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 986.678575] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1bf5b890-c994-47be-9816-30feccac7fa3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.701999] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 986.701999] env[62569]: value = "task-1250538" [ 986.701999] env[62569]: _type = "Task" [ 986.701999] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.710231] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250538, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.758557] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1ec37832-46c9-42f2-b15b-4795fdfcc977 tempest-ServersNegativeTestJSON-2046790957 tempest-ServersNegativeTestJSON-2046790957-project-member] Lock "b6e5eefc-8c06-445b-a3af-9404578b6179" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.333s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.861190] env[62569]: DEBUG nova.network.neutron [req-2b1a6184-5539-4836-8ca9-b9d6b746c505 req-618d6fca-7740-4ed3-bf18-b05501bd7546 service nova] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Updated VIF entry in instance network info cache for port 2e9fac89-877b-4c36-808d-21c6f01c992d. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 986.861633] env[62569]: DEBUG nova.network.neutron [req-2b1a6184-5539-4836-8ca9-b9d6b746c505 req-618d6fca-7740-4ed3-bf18-b05501bd7546 service nova] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Updating instance_info_cache with network_info: [{"id": "2e9fac89-877b-4c36-808d-21c6f01c992d", "address": "fa:16:3e:f3:3e:ee", "network": {"id": "73711c83-88d7-490c-bc06-44e6ed11339c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1710731049-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "44464561dec24f209ec4493ac856679b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "05b1253d-2b87-4158-9ff1-dafcf829f11f", "external-id": "nsx-vlan-transportzone-55", "segmentation_id": 55, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e9fac89-87", "ovs_interfaceid": "2e9fac89-877b-4c36-808d-21c6f01c992d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.017670] env[62569]: DEBUG nova.compute.manager [req-49cb73f4-34f5-479b-b1f3-bbbeec931e0f req-8cc54404-880f-4292-bc3d-808e5537a326 service nova] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Received event network-vif-deleted-95f1fe41-3f4f-44ce-adb6-7639e7268c5d {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 987.094054] env[62569]: INFO nova.compute.manager [-] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Took 1.39 seconds to deallocate network for instance. [ 987.121775] env[62569]: DEBUG oslo_vmware.api [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250527, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.164564] env[62569]: DEBUG oslo_vmware.api [None req-ca86470d-1e63-402f-bc09-a2adddc7d62a tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250535, 'name': ReconfigVM_Task, 'duration_secs': 0.825381} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.164891] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca86470d-1e63-402f-bc09-a2adddc7d62a tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Reconfigured VM instance instance-0000005a to attach disk [datastore1] volume-457ddfd9-7527-4e8c-8e6e-7428d784f6d3/volume-457ddfd9-7527-4e8c-8e6e-7428d784f6d3.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 987.169628] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37ce4f12-d25d-4207-958b-529129952a83 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.185751] env[62569]: DEBUG oslo_vmware.api [None req-ca86470d-1e63-402f-bc09-a2adddc7d62a tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 987.185751] env[62569]: value = "task-1250539" [ 987.185751] env[62569]: _type = "Task" [ 987.185751] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.194096] env[62569]: DEBUG oslo_vmware.api [None req-ca86470d-1e63-402f-bc09-a2adddc7d62a tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250539, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.211544] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250538, 'name': CreateVM_Task, 'duration_secs': 0.341722} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.211722] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 987.212467] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.212710] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.213064] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 987.213337] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f400923-9b49-4629-b6e7-ea505352c083 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.217860] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 987.217860] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527bb274-1439-9638-c2bf-3a156148de2c" [ 987.217860] env[62569]: _type = "Task" [ 987.217860] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.225492] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527bb274-1439-9638-c2bf-3a156148de2c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.237621] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cd9fec7f-d7ff-4679-b07a-d6288eb3aeac tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.238724] env[62569]: DEBUG oslo_concurrency.lockutils [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.037s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.240218] env[62569]: INFO nova.compute.claims [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 987.365115] env[62569]: DEBUG oslo_concurrency.lockutils [req-2b1a6184-5539-4836-8ca9-b9d6b746c505 req-618d6fca-7740-4ed3-bf18-b05501bd7546 service nova] Releasing lock "refresh_cache-9b5a1451-af46-4cd8-8e7a-f4560d8499e6" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.504076] env[62569]: DEBUG nova.compute.manager [req-545683f6-4d64-42b0-8bfa-53be4264fe1a req-cd836350-2af8-4487-9ffd-dc671c4c332c service nova] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Received event network-vif-plugged-69a7863e-3e3c-4910-be84-9e299a18442e {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 987.504317] env[62569]: DEBUG oslo_concurrency.lockutils [req-545683f6-4d64-42b0-8bfa-53be4264fe1a req-cd836350-2af8-4487-9ffd-dc671c4c332c service nova] Acquiring lock "2ce7e7ab-8a07-468f-9f9d-643d814981bd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.504525] env[62569]: DEBUG oslo_concurrency.lockutils [req-545683f6-4d64-42b0-8bfa-53be4264fe1a req-cd836350-2af8-4487-9ffd-dc671c4c332c service nova] Lock "2ce7e7ab-8a07-468f-9f9d-643d814981bd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.504697] env[62569]: DEBUG oslo_concurrency.lockutils [req-545683f6-4d64-42b0-8bfa-53be4264fe1a req-cd836350-2af8-4487-9ffd-dc671c4c332c service nova] Lock "2ce7e7ab-8a07-468f-9f9d-643d814981bd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.504872] env[62569]: DEBUG nova.compute.manager [req-545683f6-4d64-42b0-8bfa-53be4264fe1a req-cd836350-2af8-4487-9ffd-dc671c4c332c service nova] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] No waiting events found dispatching network-vif-plugged-69a7863e-3e3c-4910-be84-9e299a18442e {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 987.505311] env[62569]: WARNING nova.compute.manager [req-545683f6-4d64-42b0-8bfa-53be4264fe1a req-cd836350-2af8-4487-9ffd-dc671c4c332c service nova] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Received unexpected event network-vif-plugged-69a7863e-3e3c-4910-be84-9e299a18442e for instance with vm_state building and task_state spawning. [ 987.600088] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.617576] env[62569]: DEBUG nova.network.neutron [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Successfully updated port: 69a7863e-3e3c-4910-be84-9e299a18442e {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 987.624381] env[62569]: DEBUG oslo_vmware.api [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250527, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.696485] env[62569]: DEBUG oslo_vmware.api [None req-ca86470d-1e63-402f-bc09-a2adddc7d62a tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250539, 'name': ReconfigVM_Task, 'duration_secs': 0.19138} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.696819] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca86470d-1e63-402f-bc09-a2adddc7d62a tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269492', 'volume_id': '457ddfd9-7527-4e8c-8e6e-7428d784f6d3', 'name': 'volume-457ddfd9-7527-4e8c-8e6e-7428d784f6d3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '98089a34-074b-4bdb-92ae-f9e23d2551f1', 'attached_at': '', 'detached_at': '', 'volume_id': '457ddfd9-7527-4e8c-8e6e-7428d784f6d3', 'serial': '457ddfd9-7527-4e8c-8e6e-7428d784f6d3'} {{(pid=62569) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 987.729338] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527bb274-1439-9638-c2bf-3a156148de2c, 'name': SearchDatastore_Task, 'duration_secs': 0.010882} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.729716] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.729954] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 987.730200] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.730344] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.730534] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 987.730803] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1213dc3b-eed8-4c94-9fe5-63cdc60fed13 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.739722] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 987.739901] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 987.740623] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dee0aa68-344b-47f1-98de-1ce1bb0a144a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.749032] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 987.749032] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b1c1f4-949b-b360-9f68-903c4dd572f1" [ 987.749032] env[62569]: _type = "Task" [ 987.749032] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.756968] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b1c1f4-949b-b360-9f68-903c4dd572f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.122977] env[62569]: DEBUG oslo_concurrency.lockutils [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "refresh_cache-2ce7e7ab-8a07-468f-9f9d-643d814981bd" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.123275] env[62569]: DEBUG oslo_concurrency.lockutils [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired lock "refresh_cache-2ce7e7ab-8a07-468f-9f9d-643d814981bd" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.123384] env[62569]: DEBUG nova.network.neutron [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 988.124549] env[62569]: DEBUG oslo_vmware.api [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250527, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.260776] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b1c1f4-949b-b360-9f68-903c4dd572f1, 'name': SearchDatastore_Task, 'duration_secs': 0.008305} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.261701] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9dc65ff4-16e8-48fa-9a56-f308518cb6d8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.268344] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 988.268344] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520704d2-83bc-d71f-7ca3-8cd76db7ed43" [ 988.268344] env[62569]: _type = "Task" [ 988.268344] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.276613] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520704d2-83bc-d71f-7ca3-8cd76db7ed43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.436795] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c46b2e60-790a-41a1-9a22-9be1c6785ab7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.444333] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6585309-37b4-4dcc-b49d-7bcbfb51cd36 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.475435] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-770a6d2d-7cb9-41a0-a6e3-f9a81c3e5c37 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.483242] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b3d3fe-5f18-459c-8122-8b685b8841fc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.496279] env[62569]: DEBUG nova.compute.provider_tree [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 988.624975] env[62569]: DEBUG oslo_vmware.api [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250527, 'name': ReconfigVM_Task, 'duration_secs': 5.76819} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.625260] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.625479] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Reconfigured VM to detach interface {{(pid=62569) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 988.656383] env[62569]: DEBUG nova.network.neutron [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 988.736865] env[62569]: DEBUG nova.objects.instance [None req-ca86470d-1e63-402f-bc09-a2adddc7d62a tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lazy-loading 'flavor' on Instance uuid 98089a34-074b-4bdb-92ae-f9e23d2551f1 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 988.779120] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520704d2-83bc-d71f-7ca3-8cd76db7ed43, 'name': SearchDatastore_Task, 'duration_secs': 0.008871} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.779490] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.779848] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 9b5a1451-af46-4cd8-8e7a-f4560d8499e6/9b5a1451-af46-4cd8-8e7a-f4560d8499e6.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 988.780193] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-562d9f1a-633a-45bf-a4d7-c0bfc8f475ec {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.786913] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 988.786913] env[62569]: value = "task-1250540" [ 988.786913] env[62569]: _type = "Task" [ 988.786913] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.794973] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250540, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.802369] env[62569]: DEBUG nova.network.neutron [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Updating instance_info_cache with network_info: [{"id": "69a7863e-3e3c-4910-be84-9e299a18442e", "address": "fa:16:3e:42:22:83", "network": {"id": "884d3248-07f1-4cf7-8ceb-b94b34622384", "bridge": "br-int", "label": "tempest-ServersTestJSON-1882023778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2c6a4814c234d53bbd4844cd482fbae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69a7863e-3e", "ovs_interfaceid": "69a7863e-3e3c-4910-be84-9e299a18442e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.999852] env[62569]: DEBUG nova.scheduler.client.report [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 989.247692] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ca86470d-1e63-402f-bc09-a2adddc7d62a tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.812s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.300100] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250540, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.445046} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.303015] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 9b5a1451-af46-4cd8-8e7a-f4560d8499e6/9b5a1451-af46-4cd8-8e7a-f4560d8499e6.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 989.303462] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 989.304148] env[62569]: DEBUG oslo_concurrency.lockutils [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Releasing lock "refresh_cache-2ce7e7ab-8a07-468f-9f9d-643d814981bd" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.304549] env[62569]: DEBUG nova.compute.manager [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Instance network_info: |[{"id": "69a7863e-3e3c-4910-be84-9e299a18442e", "address": "fa:16:3e:42:22:83", "network": {"id": "884d3248-07f1-4cf7-8ceb-b94b34622384", "bridge": "br-int", "label": "tempest-ServersTestJSON-1882023778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2c6a4814c234d53bbd4844cd482fbae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69a7863e-3e", "ovs_interfaceid": "69a7863e-3e3c-4910-be84-9e299a18442e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 989.305841] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-eb744780-d02f-4e57-b6fa-497a78252755 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.310113] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:42:22:83', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e9fa4744-8702-4973-b911-ee18192a3e4b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '69a7863e-3e3c-4910-be84-9e299a18442e', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 989.318370] env[62569]: DEBUG oslo.service.loopingcall [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 989.318720] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 989.319470] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dfa55750-bea8-4cc3-b09e-350d72aa7f6f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.338825] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 989.338825] env[62569]: value = "task-1250541" [ 989.338825] env[62569]: _type = "Task" [ 989.338825] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.343829] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 989.343829] env[62569]: value = "task-1250542" [ 989.343829] env[62569]: _type = "Task" [ 989.343829] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.349492] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250541, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.354126] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250542, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.506592] env[62569]: DEBUG oslo_concurrency.lockutils [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.268s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.507191] env[62569]: DEBUG nova.compute.manager [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 989.510431] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.501s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.510630] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.512981] env[62569]: DEBUG oslo_concurrency.lockutils [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.073s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.514221] env[62569]: DEBUG oslo_concurrency.lockutils [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.516134] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.358s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.516134] env[62569]: DEBUG nova.objects.instance [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lazy-loading 'resources' on Instance uuid bcc7f73b-083b-477e-bceb-4f984a95d219 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 989.537569] env[62569]: DEBUG nova.compute.manager [req-f0d826ad-4d40-4b7e-b095-e624b1483535 req-d1d2fa6d-99ed-4ec0-aadd-67e88b863b98 service nova] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Received event network-changed-69a7863e-3e3c-4910-be84-9e299a18442e {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 989.537791] env[62569]: DEBUG nova.compute.manager [req-f0d826ad-4d40-4b7e-b095-e624b1483535 req-d1d2fa6d-99ed-4ec0-aadd-67e88b863b98 service nova] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Refreshing instance network info cache due to event network-changed-69a7863e-3e3c-4910-be84-9e299a18442e. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 989.538018] env[62569]: DEBUG oslo_concurrency.lockutils [req-f0d826ad-4d40-4b7e-b095-e624b1483535 req-d1d2fa6d-99ed-4ec0-aadd-67e88b863b98 service nova] Acquiring lock "refresh_cache-2ce7e7ab-8a07-468f-9f9d-643d814981bd" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.538199] env[62569]: DEBUG oslo_concurrency.lockutils [req-f0d826ad-4d40-4b7e-b095-e624b1483535 req-d1d2fa6d-99ed-4ec0-aadd-67e88b863b98 service nova] Acquired lock "refresh_cache-2ce7e7ab-8a07-468f-9f9d-643d814981bd" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.538412] env[62569]: DEBUG nova.network.neutron [req-f0d826ad-4d40-4b7e-b095-e624b1483535 req-d1d2fa6d-99ed-4ec0-aadd-67e88b863b98 service nova] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Refreshing network info cache for port 69a7863e-3e3c-4910-be84-9e299a18442e {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 989.542765] env[62569]: INFO nova.scheduler.client.report [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Deleted allocations for instance 09eeb937-aabc-418d-9fd3-3c7a72f4ef75 [ 989.547103] env[62569]: INFO nova.scheduler.client.report [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Deleted allocations for instance b47cd2d7-0cd2-41af-8ed1-a6dfca323516 [ 989.653246] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.653544] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.653781] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "93eb1a09-269a-44c8-8f9d-3a2b7e0d715e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.653984] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "93eb1a09-269a-44c8-8f9d-3a2b7e0d715e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.654332] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "93eb1a09-269a-44c8-8f9d-3a2b7e0d715e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.656359] env[62569]: INFO nova.compute.manager [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Terminating instance [ 989.850239] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250541, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063498} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.851021] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 989.851840] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94dbd0a9-a884-40db-805b-e660c2fdf14f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.857031] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250542, 'name': CreateVM_Task, 'duration_secs': 0.480884} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.857521] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 989.858189] env[62569]: DEBUG oslo_concurrency.lockutils [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.858356] env[62569]: DEBUG oslo_concurrency.lockutils [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.858680] env[62569]: DEBUG oslo_concurrency.lockutils [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 989.858962] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c0ce0c5-7ceb-4cfb-a8f2-29c15e4e2162 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.877856] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 9b5a1451-af46-4cd8-8e7a-f4560d8499e6/9b5a1451-af46-4cd8-8e7a-f4560d8499e6.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 989.878468] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cca3868d-a89a-4775-8257-2951f87a24f0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.894105] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 989.894105] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52dc0fe4-e41a-6691-f605-4af50db50188" [ 989.894105] env[62569]: _type = "Task" [ 989.894105] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.899743] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 989.899743] env[62569]: value = "task-1250543" [ 989.899743] env[62569]: _type = "Task" [ 989.899743] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.907628] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52dc0fe4-e41a-6691-f605-4af50db50188, 'name': SearchDatastore_Task, 'duration_secs': 0.008873} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.908209] env[62569]: DEBUG oslo_concurrency.lockutils [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.908449] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 989.908682] env[62569]: DEBUG oslo_concurrency.lockutils [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.908834] env[62569]: DEBUG oslo_concurrency.lockutils [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.909080] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 989.909265] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f392a0c4-d499-4697-b705-057fcb59e13e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.913952] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250543, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.923724] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 989.923877] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 989.924893] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3bfd97a-9c11-4612-9287-cd75cae7688d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.930073] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 989.930073] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52437274-3ae3-f7ee-9b4c-8e4977d45578" [ 989.930073] env[62569]: _type = "Task" [ 989.930073] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.937284] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52437274-3ae3-f7ee-9b4c-8e4977d45578, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.019520] env[62569]: DEBUG nova.compute.utils [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 990.024224] env[62569]: DEBUG nova.compute.manager [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 990.024404] env[62569]: DEBUG nova.network.neutron [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 990.029499] env[62569]: DEBUG oslo_concurrency.lockutils [None req-382ceb05-246e-4ed9-b315-3308d922a363 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.029810] env[62569]: DEBUG oslo_concurrency.lockutils [None req-382ceb05-246e-4ed9-b315-3308d922a363 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.053914] env[62569]: DEBUG oslo_concurrency.lockutils [None req-353be0c1-23f9-4840-aa54-786692de8ab3 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "09eeb937-aabc-418d-9fd3-3c7a72f4ef75" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.422s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.059714] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.059714] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquired lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.059714] env[62569]: DEBUG nova.network.neutron [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 990.063371] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc3a0416-f09b-4f84-a22e-b8ec3ffdff75 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "b47cd2d7-0cd2-41af-8ed1-a6dfca323516" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.657s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.084554] env[62569]: DEBUG nova.policy [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1418d7ccc2a4324aeb586f89c2f817b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4ba7e96d226942bd99ab29fe703c6fea', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 990.161071] env[62569]: DEBUG nova.compute.manager [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 990.161333] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 990.162318] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd8a0978-09ce-4c34-b71e-007603e595bb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.171805] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 990.172094] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-79e3ab56-aae5-4a72-b3b3-bf315d9cce7c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.179639] env[62569]: DEBUG oslo_vmware.api [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 990.179639] env[62569]: value = "task-1250544" [ 990.179639] env[62569]: _type = "Task" [ 990.179639] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.190810] env[62569]: DEBUG oslo_vmware.api [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250544, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.224138] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95f62c19-343d-4f6c-a89d-197fbf8c61bb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.231723] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2258d81f-087e-4737-830a-4f3e2997d516 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.265520] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a2a1289-a8c6-4e62-879d-08928a284408 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.274063] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-150b612e-f2f5-4ab7-a2b9-0dd74a5e32fb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.289703] env[62569]: DEBUG nova.compute.provider_tree [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.306349] env[62569]: DEBUG nova.network.neutron [req-f0d826ad-4d40-4b7e-b095-e624b1483535 req-d1d2fa6d-99ed-4ec0-aadd-67e88b863b98 service nova] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Updated VIF entry in instance network info cache for port 69a7863e-3e3c-4910-be84-9e299a18442e. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 990.306718] env[62569]: DEBUG nova.network.neutron [req-f0d826ad-4d40-4b7e-b095-e624b1483535 req-d1d2fa6d-99ed-4ec0-aadd-67e88b863b98 service nova] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Updating instance_info_cache with network_info: [{"id": "69a7863e-3e3c-4910-be84-9e299a18442e", "address": "fa:16:3e:42:22:83", "network": {"id": "884d3248-07f1-4cf7-8ceb-b94b34622384", "bridge": "br-int", "label": "tempest-ServersTestJSON-1882023778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2c6a4814c234d53bbd4844cd482fbae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69a7863e-3e", "ovs_interfaceid": "69a7863e-3e3c-4910-be84-9e299a18442e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.410180] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250543, 'name': ReconfigVM_Task, 'duration_secs': 0.265898} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.410485] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 9b5a1451-af46-4cd8-8e7a-f4560d8499e6/9b5a1451-af46-4cd8-8e7a-f4560d8499e6.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 990.411146] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-02a388d8-8946-44ea-b197-89891ba0d685 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.417699] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 990.417699] env[62569]: value = "task-1250545" [ 990.417699] env[62569]: _type = "Task" [ 990.417699] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.426580] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250545, 'name': Rename_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.438950] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52437274-3ae3-f7ee-9b4c-8e4977d45578, 'name': SearchDatastore_Task, 'duration_secs': 0.008122} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.439720] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a36b98e-c4cb-4e6a-b3ff-ae2a2df419b2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.445042] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 990.445042] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f6f7e1-c3b3-5f0e-0029-3b5b835cbc3a" [ 990.445042] env[62569]: _type = "Task" [ 990.445042] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.453940] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f6f7e1-c3b3-5f0e-0029-3b5b835cbc3a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.528124] env[62569]: DEBUG nova.compute.manager [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 990.534135] env[62569]: INFO nova.compute.manager [None req-382ceb05-246e-4ed9-b315-3308d922a363 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Detaching volume 4494e8a8-22d4-4334-a304-ace87e30e8ed [ 990.573219] env[62569]: INFO nova.virt.block_device [None req-382ceb05-246e-4ed9-b315-3308d922a363 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Attempting to driver detach volume 4494e8a8-22d4-4334-a304-ace87e30e8ed from mountpoint /dev/sdb [ 990.573485] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-382ceb05-246e-4ed9-b315-3308d922a363 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Volume detach. Driver type: vmdk {{(pid=62569) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 990.573696] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-382ceb05-246e-4ed9-b315-3308d922a363 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269488', 'volume_id': '4494e8a8-22d4-4334-a304-ace87e30e8ed', 'name': 'volume-4494e8a8-22d4-4334-a304-ace87e30e8ed', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '98089a34-074b-4bdb-92ae-f9e23d2551f1', 'attached_at': '', 'detached_at': '', 'volume_id': '4494e8a8-22d4-4334-a304-ace87e30e8ed', 'serial': '4494e8a8-22d4-4334-a304-ace87e30e8ed'} {{(pid=62569) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 990.574609] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7f74cb3-cb8c-4186-bfc1-578fadfe1257 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.601826] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cedbee7d-52bf-4d06-908e-cfc0619718e5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.614211] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-995351e4-e511-49ff-8f41-4c9723278ca2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.642282] env[62569]: DEBUG nova.network.neutron [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Successfully created port: cc6706bd-0644-4c6c-a181-1b01d3b5f98a {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 990.645194] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b1d1c4b-dd75-48e9-a3a7-f3406cfb61f5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.665969] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-382ceb05-246e-4ed9-b315-3308d922a363 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] The volume has not been displaced from its original location: [datastore2] volume-4494e8a8-22d4-4334-a304-ace87e30e8ed/volume-4494e8a8-22d4-4334-a304-ace87e30e8ed.vmdk. No consolidation needed. {{(pid=62569) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 990.673251] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-382ceb05-246e-4ed9-b315-3308d922a363 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Reconfiguring VM instance instance-0000005a to detach disk 2001 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 990.673633] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6454b690-9559-4d9c-a01a-b503cb29f63d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.698046] env[62569]: DEBUG oslo_vmware.api [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250544, 'name': PowerOffVM_Task, 'duration_secs': 0.19194} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.698303] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 990.698489] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 990.698821] env[62569]: DEBUG oslo_vmware.api [None req-382ceb05-246e-4ed9-b315-3308d922a363 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 990.698821] env[62569]: value = "task-1250546" [ 990.698821] env[62569]: _type = "Task" [ 990.698821] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.699045] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7c409904-23f1-4184-8dcb-2d7c01e8a86b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.709574] env[62569]: DEBUG oslo_vmware.api [None req-382ceb05-246e-4ed9-b315-3308d922a363 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250546, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.770227] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 990.770468] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 990.770699] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Deleting the datastore file [datastore1] 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 990.770980] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e42cff18-6bcf-45d7-8679-3d96e7475d61 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.777815] env[62569]: DEBUG oslo_vmware.api [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 990.777815] env[62569]: value = "task-1250548" [ 990.777815] env[62569]: _type = "Task" [ 990.777815] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.789552] env[62569]: DEBUG oslo_vmware.api [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250548, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.792542] env[62569]: DEBUG nova.scheduler.client.report [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 990.809771] env[62569]: DEBUG oslo_concurrency.lockutils [req-f0d826ad-4d40-4b7e-b095-e624b1483535 req-d1d2fa6d-99ed-4ec0-aadd-67e88b863b98 service nova] Releasing lock "refresh_cache-2ce7e7ab-8a07-468f-9f9d-643d814981bd" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.857791] env[62569]: INFO nova.network.neutron [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Port 0d27f2d0-6f98-4031-97c8-33df605f1bb0 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 990.858183] env[62569]: DEBUG nova.network.neutron [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Updating instance_info_cache with network_info: [{"id": "6ca4b28a-1255-4f88-a7b6-54b0378768b1", "address": "fa:16:3e:03:e7:ad", "network": {"id": "29fcbfe8-35ce-4e8d-874b-2b14ba720adf", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-719645861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "530ca7e7924743ab91a362a064a3111b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ca4b28a-12", "ovs_interfaceid": "6ca4b28a-1255-4f88-a7b6-54b0378768b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.930479] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250545, 'name': Rename_Task, 'duration_secs': 0.145542} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.930820] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 990.931118] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a19f9dee-bb5f-4607-9997-b5df402af895 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.937656] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 990.937656] env[62569]: value = "task-1250549" [ 990.937656] env[62569]: _type = "Task" [ 990.937656] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.945271] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250549, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.952932] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52f6f7e1-c3b3-5f0e-0029-3b5b835cbc3a, 'name': SearchDatastore_Task, 'duration_secs': 0.009512} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.953245] env[62569]: DEBUG oslo_concurrency.lockutils [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.953552] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 2ce7e7ab-8a07-468f-9f9d-643d814981bd/2ce7e7ab-8a07-468f-9f9d-643d814981bd.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 990.953849] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-633398b2-2408-47d3-bcae-1bae0ea9acae {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.960297] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 990.960297] env[62569]: value = "task-1250550" [ 990.960297] env[62569]: _type = "Task" [ 990.960297] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.967622] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250550, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.212498] env[62569]: DEBUG oslo_vmware.api [None req-382ceb05-246e-4ed9-b315-3308d922a363 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250546, 'name': ReconfigVM_Task, 'duration_secs': 0.243172} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.212810] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-382ceb05-246e-4ed9-b315-3308d922a363 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Reconfigured VM instance instance-0000005a to detach disk 2001 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 991.217560] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f330f428-6c06-4752-85ec-06075db719b2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.235523] env[62569]: DEBUG oslo_vmware.api [None req-382ceb05-246e-4ed9-b315-3308d922a363 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 991.235523] env[62569]: value = "task-1250551" [ 991.235523] env[62569]: _type = "Task" [ 991.235523] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.244147] env[62569]: DEBUG oslo_vmware.api [None req-382ceb05-246e-4ed9-b315-3308d922a363 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250551, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.272193] env[62569]: DEBUG oslo_concurrency.lockutils [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.272537] env[62569]: DEBUG oslo_concurrency.lockutils [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.288120] env[62569]: DEBUG oslo_vmware.api [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250548, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.223408} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.288371] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 991.288555] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 991.288783] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 991.288973] env[62569]: INFO nova.compute.manager [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Took 1.13 seconds to destroy the instance on the hypervisor. [ 991.289225] env[62569]: DEBUG oslo.service.loopingcall [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 991.289427] env[62569]: DEBUG nova.compute.manager [-] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 991.289516] env[62569]: DEBUG nova.network.neutron [-] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 991.298007] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.782s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.303016] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 5.013s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.327295] env[62569]: INFO nova.scheduler.client.report [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Deleted allocations for instance bcc7f73b-083b-477e-bceb-4f984a95d219 [ 991.361113] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Releasing lock "refresh_cache-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.450972] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250549, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.472458] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250550, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.537218] env[62569]: DEBUG nova.compute.manager [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 991.573239] env[62569]: DEBUG nova.virt.hardware [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 991.573519] env[62569]: DEBUG nova.virt.hardware [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 991.573760] env[62569]: DEBUG nova.virt.hardware [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 991.574092] env[62569]: DEBUG nova.virt.hardware [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 991.574181] env[62569]: DEBUG nova.virt.hardware [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 991.574320] env[62569]: DEBUG nova.virt.hardware [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 991.574535] env[62569]: DEBUG nova.virt.hardware [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 991.574714] env[62569]: DEBUG nova.virt.hardware [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 991.574882] env[62569]: DEBUG nova.virt.hardware [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 991.575171] env[62569]: DEBUG nova.virt.hardware [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 991.575327] env[62569]: DEBUG nova.virt.hardware [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 991.576304] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88c2021a-8d0f-49f4-a729-bae2f3371b0b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.591534] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f1f3631-9347-4413-b7d7-f828ea772989 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.745518] env[62569]: DEBUG oslo_vmware.api [None req-382ceb05-246e-4ed9-b315-3308d922a363 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250551, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.775231] env[62569]: DEBUG nova.compute.manager [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 991.836060] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b8f44cf7-88e5-4ecd-a1fd-d9f71c15565e tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "bcc7f73b-083b-477e-bceb-4f984a95d219" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.416s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.867631] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47b68134-0be7-499d-9ce2-0cc16a0ece32 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "interface-93eb1a09-269a-44c8-8f9d-3a2b7e0d715e-0d27f2d0-6f98-4031-97c8-33df605f1bb0" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.829s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.949049] env[62569]: DEBUG oslo_vmware.api [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250549, 'name': PowerOnVM_Task, 'duration_secs': 0.915105} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.949338] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 991.949576] env[62569]: INFO nova.compute.manager [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Took 7.81 seconds to spawn the instance on the hypervisor. [ 991.949761] env[62569]: DEBUG nova.compute.manager [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 991.950616] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b71c9b34-51d9-4127-a812-4be610718863 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.970327] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250550, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.647816} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.970562] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 2ce7e7ab-8a07-468f-9f9d-643d814981bd/2ce7e7ab-8a07-468f-9f9d-643d814981bd.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 991.970768] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 991.971028] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e76d28a0-d3c0-4b2b-9899-9642b08eddc1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.977141] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 991.977141] env[62569]: value = "task-1250552" [ 991.977141] env[62569]: _type = "Task" [ 991.977141] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.985421] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250552, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.028874] env[62569]: DEBUG nova.compute.manager [req-ba1037d8-ad42-4d50-a01e-1f4236a1abe0 req-3c5236a8-1e6d-4698-ae52-3d8d02dfbcf5 service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Received event network-vif-deleted-6ca4b28a-1255-4f88-a7b6-54b0378768b1 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 992.029515] env[62569]: INFO nova.compute.manager [req-ba1037d8-ad42-4d50-a01e-1f4236a1abe0 req-3c5236a8-1e6d-4698-ae52-3d8d02dfbcf5 service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Neutron deleted interface 6ca4b28a-1255-4f88-a7b6-54b0378768b1; detaching it from the instance and deleting it from the info cache [ 992.029515] env[62569]: DEBUG nova.network.neutron [req-ba1037d8-ad42-4d50-a01e-1f4236a1abe0 req-3c5236a8-1e6d-4698-ae52-3d8d02dfbcf5 service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.113318] env[62569]: DEBUG nova.compute.manager [req-bebce3c9-ed24-4a23-bc3c-e19c820a11a7 req-afd8cc2f-b962-4d4d-8960-5bbeced3e62f service nova] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Received event network-vif-plugged-cc6706bd-0644-4c6c-a181-1b01d3b5f98a {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 992.113516] env[62569]: DEBUG oslo_concurrency.lockutils [req-bebce3c9-ed24-4a23-bc3c-e19c820a11a7 req-afd8cc2f-b962-4d4d-8960-5bbeced3e62f service nova] Acquiring lock "4df2078e-d7ce-4a3f-904b-f3303071ffff-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.113736] env[62569]: DEBUG oslo_concurrency.lockutils [req-bebce3c9-ed24-4a23-bc3c-e19c820a11a7 req-afd8cc2f-b962-4d4d-8960-5bbeced3e62f service nova] Lock "4df2078e-d7ce-4a3f-904b-f3303071ffff-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.113943] env[62569]: DEBUG oslo_concurrency.lockutils [req-bebce3c9-ed24-4a23-bc3c-e19c820a11a7 req-afd8cc2f-b962-4d4d-8960-5bbeced3e62f service nova] Lock "4df2078e-d7ce-4a3f-904b-f3303071ffff-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.114406] env[62569]: DEBUG nova.compute.manager [req-bebce3c9-ed24-4a23-bc3c-e19c820a11a7 req-afd8cc2f-b962-4d4d-8960-5bbeced3e62f service nova] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] No waiting events found dispatching network-vif-plugged-cc6706bd-0644-4c6c-a181-1b01d3b5f98a {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 992.114617] env[62569]: WARNING nova.compute.manager [req-bebce3c9-ed24-4a23-bc3c-e19c820a11a7 req-afd8cc2f-b962-4d4d-8960-5bbeced3e62f service nova] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Received unexpected event network-vif-plugged-cc6706bd-0644-4c6c-a181-1b01d3b5f98a for instance with vm_state building and task_state spawning. [ 992.212765] env[62569]: DEBUG nova.network.neutron [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Successfully updated port: cc6706bd-0644-4c6c-a181-1b01d3b5f98a {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 992.248693] env[62569]: DEBUG oslo_vmware.api [None req-382ceb05-246e-4ed9-b315-3308d922a363 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250551, 'name': ReconfigVM_Task, 'duration_secs': 0.86147} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.249030] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-382ceb05-246e-4ed9-b315-3308d922a363 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269488', 'volume_id': '4494e8a8-22d4-4334-a304-ace87e30e8ed', 'name': 'volume-4494e8a8-22d4-4334-a304-ace87e30e8ed', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '98089a34-074b-4bdb-92ae-f9e23d2551f1', 'attached_at': '', 'detached_at': '', 'volume_id': '4494e8a8-22d4-4334-a304-ace87e30e8ed', 'serial': '4494e8a8-22d4-4334-a304-ace87e30e8ed'} {{(pid=62569) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 992.308867] env[62569]: DEBUG oslo_concurrency.lockutils [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.338365] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance e4eadcdf-a04a-4255-ba1c-fe20156c655f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.338594] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 534c9ece-c8df-4528-83d8-c6c0ef8c4793 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.338787] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 188cac03-4034-4a02-973a-fb1906399fd1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.338972] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 98089a34-074b-4bdb-92ae-f9e23d2551f1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.339175] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.339389] env[62569]: WARNING nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 35350164-e295-40ea-85a6-e8e79c630e0f is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 992.339786] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance d56faf39-6caf-4489-98d7-342fc5fb7d40 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.339786] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 9b5a1451-af46-4cd8-8e7a-f4560d8499e6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.339993] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 2ce7e7ab-8a07-468f-9f9d-643d814981bd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.340111] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 4df2078e-d7ce-4a3f-904b-f3303071ffff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 992.468036] env[62569]: INFO nova.compute.manager [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Took 21.40 seconds to build instance. [ 992.494327] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250552, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074516} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.494676] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 992.495549] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-842d3604-3a51-45d3-befc-92238488b11f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.513989] env[62569]: DEBUG nova.network.neutron [-] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.523161] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 2ce7e7ab-8a07-468f-9f9d-643d814981bd/2ce7e7ab-8a07-468f-9f9d-643d814981bd.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 992.523884] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b250881-a3de-431f-a399-2dd67b5d4ce1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.538712] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-34ab6696-f5b5-4e58-a27d-64303cabb58a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.549242] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e041dd12-3f02-4601-98ed-58fe897f7205 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.559667] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 992.559667] env[62569]: value = "task-1250553" [ 992.559667] env[62569]: _type = "Task" [ 992.559667] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.568923] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250553, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.578942] env[62569]: DEBUG nova.compute.manager [req-ba1037d8-ad42-4d50-a01e-1f4236a1abe0 req-3c5236a8-1e6d-4698-ae52-3d8d02dfbcf5 service nova] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Detach interface failed, port_id=6ca4b28a-1255-4f88-a7b6-54b0378768b1, reason: Instance 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 992.715726] env[62569]: DEBUG oslo_concurrency.lockutils [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "refresh_cache-4df2078e-d7ce-4a3f-904b-f3303071ffff" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.716984] env[62569]: DEBUG oslo_concurrency.lockutils [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired lock "refresh_cache-4df2078e-d7ce-4a3f-904b-f3303071ffff" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.716984] env[62569]: DEBUG nova.network.neutron [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 992.807742] env[62569]: DEBUG nova.objects.instance [None req-382ceb05-246e-4ed9-b315-3308d922a363 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lazy-loading 'flavor' on Instance uuid 98089a34-074b-4bdb-92ae-f9e23d2551f1 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.843828] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 992.844118] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=62569) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 992.844280] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2240MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=62569) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 992.969962] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8f2a7b76-c033-4207-ab71-c96387c03fbb tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Lock "9b5a1451-af46-4cd8-8e7a-f4560d8499e6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.916s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.980812] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e851156c-b155-4f72-a2db-ee9eef32e5d2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.986620] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28ade694-dcd8-4baa-a325-44379bec0d5f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.018683] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b587af6b-e441-43fb-9181-30799ea794ab {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.026452] env[62569]: INFO nova.compute.manager [-] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Took 1.74 seconds to deallocate network for instance. [ 993.028964] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8255e2f9-7c67-486c-a02e-b73e20df4332 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.048197] env[62569]: DEBUG nova.compute.provider_tree [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 993.070343] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250553, 'name': ReconfigVM_Task, 'duration_secs': 0.323972} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.070638] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 2ce7e7ab-8a07-468f-9f9d-643d814981bd/2ce7e7ab-8a07-468f-9f9d-643d814981bd.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 993.071524] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5497abd6-ceda-48e4-82da-1ee93c05211c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.078080] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 993.078080] env[62569]: value = "task-1250554" [ 993.078080] env[62569]: _type = "Task" [ 993.078080] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.086653] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250554, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.246923] env[62569]: DEBUG nova.network.neutron [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 993.402411] env[62569]: DEBUG nova.network.neutron [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Updating instance_info_cache with network_info: [{"id": "cc6706bd-0644-4c6c-a181-1b01d3b5f98a", "address": "fa:16:3e:4f:87:64", "network": {"id": "c61b7867-7897-43ae-8f5b-19834c4390ea", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-512066189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ba7e96d226942bd99ab29fe703c6fea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc6706bd-06", "ovs_interfaceid": "cc6706bd-0644-4c6c-a181-1b01d3b5f98a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.538831] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.551623] env[62569]: INFO nova.compute.manager [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Rescuing [ 993.552024] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquiring lock "refresh_cache-9b5a1451-af46-4cd8-8e7a-f4560d8499e6" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.552203] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquired lock "refresh_cache-9b5a1451-af46-4cd8-8e7a-f4560d8499e6" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.552377] env[62569]: DEBUG nova.network.neutron [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 993.554036] env[62569]: DEBUG nova.scheduler.client.report [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 993.587594] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250554, 'name': Rename_Task, 'duration_secs': 0.153207} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.587901] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 993.588255] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fc188200-a14a-479e-8f7c-13c7b833fb8c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.594382] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 993.594382] env[62569]: value = "task-1250555" [ 993.594382] env[62569]: _type = "Task" [ 993.594382] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.601724] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250555, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.815420] env[62569]: DEBUG oslo_concurrency.lockutils [None req-382ceb05-246e-4ed9-b315-3308d922a363 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.785s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.905565] env[62569]: DEBUG oslo_concurrency.lockutils [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Releasing lock "refresh_cache-4df2078e-d7ce-4a3f-904b-f3303071ffff" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.905565] env[62569]: DEBUG nova.compute.manager [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Instance network_info: |[{"id": "cc6706bd-0644-4c6c-a181-1b01d3b5f98a", "address": "fa:16:3e:4f:87:64", "network": {"id": "c61b7867-7897-43ae-8f5b-19834c4390ea", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-512066189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ba7e96d226942bd99ab29fe703c6fea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc6706bd-06", "ovs_interfaceid": "cc6706bd-0644-4c6c-a181-1b01d3b5f98a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 993.905886] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4f:87:64', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6685c85e-be1e-4b7b-a6cc-3e50e59b6567', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cc6706bd-0644-4c6c-a181-1b01d3b5f98a', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 993.913316] env[62569]: DEBUG oslo.service.loopingcall [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 993.913824] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 993.914051] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-96450ad4-68c0-408a-85ac-1875f712679e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.933940] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 993.933940] env[62569]: value = "task-1250556" [ 993.933940] env[62569]: _type = "Task" [ 993.933940] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.940859] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250556, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.969490] env[62569]: DEBUG nova.compute.manager [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Stashing vm_state: active {{(pid=62569) _prep_resize /opt/stack/nova/nova/compute/manager.py:5997}} [ 994.060363] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62569) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 994.060565] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.758s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.060886] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.461s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.061099] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.062993] env[62569]: DEBUG oslo_concurrency.lockutils [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.754s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.064510] env[62569]: INFO nova.compute.claims [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 994.088652] env[62569]: INFO nova.scheduler.client.report [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Deleted allocations for instance 35350164-e295-40ea-85a6-e8e79c630e0f [ 994.104585] env[62569]: DEBUG oslo_vmware.api [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250555, 'name': PowerOnVM_Task, 'duration_secs': 0.440122} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.104902] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 994.105207] env[62569]: INFO nova.compute.manager [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Took 7.65 seconds to spawn the instance on the hypervisor. [ 994.105423] env[62569]: DEBUG nova.compute.manager [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 994.106344] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f8ec714-e595-440d-8825-9bd1e917a2d5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.138843] env[62569]: DEBUG nova.compute.manager [req-101be6e0-ebce-4715-ab0b-d93153e4574d req-bdcb3ac5-446e-474c-a442-744bbfb64310 service nova] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Received event network-changed-cc6706bd-0644-4c6c-a181-1b01d3b5f98a {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 994.138843] env[62569]: DEBUG nova.compute.manager [req-101be6e0-ebce-4715-ab0b-d93153e4574d req-bdcb3ac5-446e-474c-a442-744bbfb64310 service nova] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Refreshing instance network info cache due to event network-changed-cc6706bd-0644-4c6c-a181-1b01d3b5f98a. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 994.139154] env[62569]: DEBUG oslo_concurrency.lockutils [req-101be6e0-ebce-4715-ab0b-d93153e4574d req-bdcb3ac5-446e-474c-a442-744bbfb64310 service nova] Acquiring lock "refresh_cache-4df2078e-d7ce-4a3f-904b-f3303071ffff" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.139154] env[62569]: DEBUG oslo_concurrency.lockutils [req-101be6e0-ebce-4715-ab0b-d93153e4574d req-bdcb3ac5-446e-474c-a442-744bbfb64310 service nova] Acquired lock "refresh_cache-4df2078e-d7ce-4a3f-904b-f3303071ffff" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.139260] env[62569]: DEBUG nova.network.neutron [req-101be6e0-ebce-4715-ab0b-d93153e4574d req-bdcb3ac5-446e-474c-a442-744bbfb64310 service nova] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Refreshing network info cache for port cc6706bd-0644-4c6c-a181-1b01d3b5f98a {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 994.323185] env[62569]: DEBUG nova.network.neutron [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Updating instance_info_cache with network_info: [{"id": "2e9fac89-877b-4c36-808d-21c6f01c992d", "address": "fa:16:3e:f3:3e:ee", "network": {"id": "73711c83-88d7-490c-bc06-44e6ed11339c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1710731049-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "44464561dec24f209ec4493ac856679b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "05b1253d-2b87-4158-9ff1-dafcf829f11f", "external-id": "nsx-vlan-transportzone-55", "segmentation_id": 55, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e9fac89-87", "ovs_interfaceid": "2e9fac89-877b-4c36-808d-21c6f01c992d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.335949] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c4002720-00ea-49a2-b49e-c18f700b338b tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.336207] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c4002720-00ea-49a2-b49e-c18f700b338b tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.443594] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250556, 'name': CreateVM_Task, 'duration_secs': 0.32332} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.443910] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 994.444501] env[62569]: DEBUG oslo_concurrency.lockutils [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.444672] env[62569]: DEBUG oslo_concurrency.lockutils [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.445009] env[62569]: DEBUG oslo_concurrency.lockutils [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 994.445277] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-303d1c2a-8c51-432c-9f69-e00249794eb2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.449743] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 994.449743] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a7837c-d74b-7ffc-e412-75f54a1f6cd2" [ 994.449743] env[62569]: _type = "Task" [ 994.449743] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.457229] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a7837c-d74b-7ffc-e412-75f54a1f6cd2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.486999] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.597132] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ae28ef74-b407-4323-8cb9-44432b39c202 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "35350164-e295-40ea-85a6-e8e79c630e0f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.038s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.622926] env[62569]: INFO nova.compute.manager [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Took 16.01 seconds to build instance. [ 994.826254] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Releasing lock "refresh_cache-9b5a1451-af46-4cd8-8e7a-f4560d8499e6" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.839409] env[62569]: INFO nova.compute.manager [None req-c4002720-00ea-49a2-b49e-c18f700b338b tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Detaching volume 457ddfd9-7527-4e8c-8e6e-7428d784f6d3 [ 994.885128] env[62569]: INFO nova.virt.block_device [None req-c4002720-00ea-49a2-b49e-c18f700b338b tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Attempting to driver detach volume 457ddfd9-7527-4e8c-8e6e-7428d784f6d3 from mountpoint /dev/sdc [ 994.885908] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4002720-00ea-49a2-b49e-c18f700b338b tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Volume detach. Driver type: vmdk {{(pid=62569) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 994.886205] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4002720-00ea-49a2-b49e-c18f700b338b tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269492', 'volume_id': '457ddfd9-7527-4e8c-8e6e-7428d784f6d3', 'name': 'volume-457ddfd9-7527-4e8c-8e6e-7428d784f6d3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '98089a34-074b-4bdb-92ae-f9e23d2551f1', 'attached_at': '', 'detached_at': '', 'volume_id': '457ddfd9-7527-4e8c-8e6e-7428d784f6d3', 'serial': '457ddfd9-7527-4e8c-8e6e-7428d784f6d3'} {{(pid=62569) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 994.887098] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4663ac8-cbed-46f9-8277-a49842616aa2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.912024] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b2447ae-1d39-448d-a357-5b7da6581f7a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.918730] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e000118-970c-46ba-bfd0-815585f00b09 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.941304] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30baf2ba-381a-41a8-9427-de034cd6f640 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.957209] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4002720-00ea-49a2-b49e-c18f700b338b tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] The volume has not been displaced from its original location: [datastore1] volume-457ddfd9-7527-4e8c-8e6e-7428d784f6d3/volume-457ddfd9-7527-4e8c-8e6e-7428d784f6d3.vmdk. No consolidation needed. {{(pid=62569) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 994.962828] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4002720-00ea-49a2-b49e-c18f700b338b tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Reconfiguring VM instance instance-0000005a to detach disk 2002 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 994.965985] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18031d71-10d4-450e-9a48-4281daa11245 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.988286] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a7837c-d74b-7ffc-e412-75f54a1f6cd2, 'name': SearchDatastore_Task, 'duration_secs': 0.01104} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.990009] env[62569]: DEBUG oslo_concurrency.lockutils [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.990433] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 994.990816] env[62569]: DEBUG oslo_concurrency.lockutils [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.991274] env[62569]: DEBUG oslo_concurrency.lockutils [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.991445] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 994.991920] env[62569]: DEBUG oslo_vmware.api [None req-c4002720-00ea-49a2-b49e-c18f700b338b tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 994.991920] env[62569]: value = "task-1250557" [ 994.991920] env[62569]: _type = "Task" [ 994.991920] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.992327] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-53d60b60-e66b-496b-8e6b-e3a55d3e67c3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.004295] env[62569]: DEBUG oslo_vmware.api [None req-c4002720-00ea-49a2-b49e-c18f700b338b tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250557, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.005973] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 995.006387] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 995.007061] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1bf9586f-c3a6-4d38-a4f3-c32546c785b4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.013401] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 995.013401] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5230b63e-0ea0-696c-cf3c-3685aae8b94f" [ 995.013401] env[62569]: _type = "Task" [ 995.013401] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.021843] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5230b63e-0ea0-696c-cf3c-3685aae8b94f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.026985] env[62569]: DEBUG nova.network.neutron [req-101be6e0-ebce-4715-ab0b-d93153e4574d req-bdcb3ac5-446e-474c-a442-744bbfb64310 service nova] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Updated VIF entry in instance network info cache for port cc6706bd-0644-4c6c-a181-1b01d3b5f98a. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 995.027439] env[62569]: DEBUG nova.network.neutron [req-101be6e0-ebce-4715-ab0b-d93153e4574d req-bdcb3ac5-446e-474c-a442-744bbfb64310 service nova] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Updating instance_info_cache with network_info: [{"id": "cc6706bd-0644-4c6c-a181-1b01d3b5f98a", "address": "fa:16:3e:4f:87:64", "network": {"id": "c61b7867-7897-43ae-8f5b-19834c4390ea", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-512066189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4ba7e96d226942bd99ab29fe703c6fea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6685c85e-be1e-4b7b-a6cc-3e50e59b6567", "external-id": "nsx-vlan-transportzone-129", "segmentation_id": 129, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc6706bd-06", "ovs_interfaceid": "cc6706bd-0644-4c6c-a181-1b01d3b5f98a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.125815] env[62569]: DEBUG oslo_concurrency.lockutils [None req-acbc8062-08d2-4687-9bef-61f43287540b tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "2ce7e7ab-8a07-468f-9f9d-643d814981bd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.529s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.244169] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4d87896-6e7b-4963-a26f-3728becdd10d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.253023] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2fdac92-0d2e-43a7-8080-33a26437b965 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.282438] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6f26036-1122-4128-b22e-99879cf4fde0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.290481] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64b26f13-170b-433d-b604-4da2fdde6e90 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.305979] env[62569]: DEBUG nova.compute.provider_tree [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 995.504282] env[62569]: DEBUG oslo_vmware.api [None req-c4002720-00ea-49a2-b49e-c18f700b338b tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250557, 'name': ReconfigVM_Task, 'duration_secs': 0.247161} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.504588] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4002720-00ea-49a2-b49e-c18f700b338b tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Reconfigured VM instance instance-0000005a to detach disk 2002 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 995.509408] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f7a01b9e-35ca-45eb-b11a-96ec7d787483 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.528388] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5230b63e-0ea0-696c-cf3c-3685aae8b94f, 'name': SearchDatastore_Task, 'duration_secs': 0.009418} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.530813] env[62569]: DEBUG oslo_concurrency.lockutils [req-101be6e0-ebce-4715-ab0b-d93153e4574d req-bdcb3ac5-446e-474c-a442-744bbfb64310 service nova] Releasing lock "refresh_cache-4df2078e-d7ce-4a3f-904b-f3303071ffff" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.531278] env[62569]: DEBUG oslo_vmware.api [None req-c4002720-00ea-49a2-b49e-c18f700b338b tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 995.531278] env[62569]: value = "task-1250558" [ 995.531278] env[62569]: _type = "Task" [ 995.531278] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.531446] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82b01476-e528-4d0c-b6c4-ecfb0cc6efc6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.539735] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 995.539735] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5275f7a5-e4ae-080a-be7c-b884759df9a6" [ 995.539735] env[62569]: _type = "Task" [ 995.539735] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.544076] env[62569]: DEBUG oslo_vmware.api [None req-c4002720-00ea-49a2-b49e-c18f700b338b tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250558, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.551668] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5275f7a5-e4ae-080a-be7c-b884759df9a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.658560] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "4bf2f0fb-420f-4c13-97c1-974af7c49658" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.658765] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "4bf2f0fb-420f-4c13-97c1-974af7c49658" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.811841] env[62569]: DEBUG nova.scheduler.client.report [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 996.043601] env[62569]: DEBUG oslo_vmware.api [None req-c4002720-00ea-49a2-b49e-c18f700b338b tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250558, 'name': ReconfigVM_Task, 'duration_secs': 0.147578} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.043940] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4002720-00ea-49a2-b49e-c18f700b338b tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269492', 'volume_id': '457ddfd9-7527-4e8c-8e6e-7428d784f6d3', 'name': 'volume-457ddfd9-7527-4e8c-8e6e-7428d784f6d3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '98089a34-074b-4bdb-92ae-f9e23d2551f1', 'attached_at': '', 'detached_at': '', 'volume_id': '457ddfd9-7527-4e8c-8e6e-7428d784f6d3', 'serial': '457ddfd9-7527-4e8c-8e6e-7428d784f6d3'} {{(pid=62569) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 996.054649] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5275f7a5-e4ae-080a-be7c-b884759df9a6, 'name': SearchDatastore_Task, 'duration_secs': 0.013319} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.054911] env[62569]: DEBUG oslo_concurrency.lockutils [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.055186] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 4df2078e-d7ce-4a3f-904b-f3303071ffff/4df2078e-d7ce-4a3f-904b-f3303071ffff.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 996.055440] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-007d1983-af14-4a77-88a9-e10f641a9a5a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.063061] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 996.063061] env[62569]: value = "task-1250559" [ 996.063061] env[62569]: _type = "Task" [ 996.063061] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.070104] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250559, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.142065] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "c896f38c-f313-491b-86c5-a785721667d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.142065] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "c896f38c-f313-491b-86c5-a785721667d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.164066] env[62569]: DEBUG nova.compute.manager [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 996.316695] env[62569]: DEBUG oslo_concurrency.lockutils [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.253s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.317332] env[62569]: DEBUG nova.compute.manager [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 996.320459] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.782s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.320701] env[62569]: DEBUG nova.objects.instance [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lazy-loading 'resources' on Instance uuid 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.356931] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 996.357396] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4532c6de-911c-4efa-b801-8586e97811ea {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.370638] env[62569]: DEBUG oslo_vmware.api [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 996.370638] env[62569]: value = "task-1250560" [ 996.370638] env[62569]: _type = "Task" [ 996.370638] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.382084] env[62569]: DEBUG oslo_vmware.api [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250560, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.572895] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250559, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.433012} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.573270] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 4df2078e-d7ce-4a3f-904b-f3303071ffff/4df2078e-d7ce-4a3f-904b-f3303071ffff.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 996.573479] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 996.573759] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4bad2e25-9618-490a-96bc-09ba22164d81 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.580768] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 996.580768] env[62569]: value = "task-1250561" [ 996.580768] env[62569]: _type = "Task" [ 996.580768] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.588988] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250561, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.600769] env[62569]: DEBUG nova.objects.instance [None req-c4002720-00ea-49a2-b49e-c18f700b338b tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lazy-loading 'flavor' on Instance uuid 98089a34-074b-4bdb-92ae-f9e23d2551f1 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.644224] env[62569]: DEBUG nova.compute.manager [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 996.685592] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.823863] env[62569]: DEBUG nova.compute.utils [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 996.827943] env[62569]: DEBUG nova.compute.manager [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 996.828124] env[62569]: DEBUG nova.network.neutron [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 996.872683] env[62569]: DEBUG nova.policy [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b9e0da546a03426782fcacd06a7afa8a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '00bdb4c3f88f4d61834f7e961629ef3a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 996.884952] env[62569]: DEBUG oslo_vmware.api [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250560, 'name': PowerOffVM_Task, 'duration_secs': 0.263963} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.885248] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 996.886073] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c5deae-e1f5-4aae-9f9f-856adf24b0ed {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.909198] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c88c3be4-5f5d-4672-9677-0a1c476cf2bb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.949337] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 996.949337] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d3422ad1-be7b-47d4-a0f0-4316d0347c42 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.957715] env[62569]: DEBUG oslo_vmware.api [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 996.957715] env[62569]: value = "task-1250562" [ 996.957715] env[62569]: _type = "Task" [ 996.957715] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.976056] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] VM already powered off {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 996.976383] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 996.976745] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.977073] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.977413] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 996.981591] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cb2ca9c8-13b8-4237-8ac6-156ac2d3a2b2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.990814] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 996.991390] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 996.994045] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-504b53d7-19d3-4f76-b779-1c69b34704c9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.999928] env[62569]: DEBUG oslo_vmware.api [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 996.999928] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52176db5-b1b0-cdfe-2427-68bf532be69b" [ 996.999928] env[62569]: _type = "Task" [ 996.999928] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.019623] env[62569]: DEBUG oslo_vmware.api [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52176db5-b1b0-cdfe-2427-68bf532be69b, 'name': SearchDatastore_Task, 'duration_secs': 0.008859} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.021350] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-58eed159-d902-4f11-ac93-18515f2b9a5f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.029768] env[62569]: DEBUG oslo_vmware.api [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 997.029768] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52742896-74ac-a43d-55fc-fd6a55bb6a2c" [ 997.029768] env[62569]: _type = "Task" [ 997.029768] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.045384] env[62569]: DEBUG oslo_vmware.api [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52742896-74ac-a43d-55fc-fd6a55bb6a2c, 'name': SearchDatastore_Task, 'duration_secs': 0.010233} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.045668] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.046449] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 9b5a1451-af46-4cd8-8e7a-f4560d8499e6/f77800cf-af0d-4e9c-b312-2c59488f2c7c-rescue.vmdk. {{(pid=62569) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 997.046449] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1436fcc5-a4ec-48a6-b2bc-2beb31c34fb7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.055517] env[62569]: DEBUG oslo_vmware.api [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 997.055517] env[62569]: value = "task-1250563" [ 997.055517] env[62569]: _type = "Task" [ 997.055517] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.063682] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fa0120d-9a6a-4e7f-ad34-715012a5896c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.069132] env[62569]: DEBUG oslo_vmware.api [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250563, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.073582] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40971d0d-bdb0-4861-9785-c7f080a775dd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.109685] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f75e782d-196b-454a-802d-44a3bc7deb1f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.121986] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7461c271-6d4b-4309-b7ae-79a8d82c0d99 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.125921] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250561, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06764} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.127055] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 997.127987] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fb8c658-3c8e-42f0-8dc4-7f66d2e3384f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.140220] env[62569]: DEBUG nova.compute.provider_tree [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 997.163520] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 4df2078e-d7ce-4a3f-904b-f3303071ffff/4df2078e-d7ce-4a3f-904b-f3303071ffff.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 997.166318] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f8b308d-4a51-441d-9d49-e9f3634abd43 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.188060] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 997.188060] env[62569]: value = "task-1250564" [ 997.188060] env[62569]: _type = "Task" [ 997.188060] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.192670] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.196900] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250564, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.301029] env[62569]: DEBUG nova.network.neutron [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Successfully created port: 8b78baca-9288-4ab0-b21d-7fb96baeaf42 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 997.328700] env[62569]: DEBUG nova.compute.manager [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 997.565260] env[62569]: DEBUG oslo_vmware.api [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250563, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.482409} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.565547] env[62569]: INFO nova.virt.vmwareapi.ds_util [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 9b5a1451-af46-4cd8-8e7a-f4560d8499e6/f77800cf-af0d-4e9c-b312-2c59488f2c7c-rescue.vmdk. [ 997.566419] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c4c5bdf-6e20-4f45-a4d8-b03acce63e99 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.595285] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 9b5a1451-af46-4cd8-8e7a-f4560d8499e6/f77800cf-af0d-4e9c-b312-2c59488f2c7c-rescue.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 997.595683] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1a699a3-b153-42a2-8bd0-ab9c3aa7e77e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.614376] env[62569]: DEBUG oslo_vmware.api [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 997.614376] env[62569]: value = "task-1250565" [ 997.614376] env[62569]: _type = "Task" [ 997.614376] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.614906] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c4002720-00ea-49a2-b49e-c18f700b338b tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.279s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.623887] env[62569]: DEBUG oslo_vmware.api [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250565, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.642721] env[62569]: DEBUG nova.scheduler.client.report [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 997.698566] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250564, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.124538] env[62569]: DEBUG oslo_vmware.api [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250565, 'name': ReconfigVM_Task, 'duration_secs': 0.295627} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.124874] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 9b5a1451-af46-4cd8-8e7a-f4560d8499e6/f77800cf-af0d-4e9c-b312-2c59488f2c7c-rescue.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 998.125727] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27c1d2d7-a7c4-4697-bed6-2a6f46a28d37 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.148955] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.828s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.156180] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.669s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.157615] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71da00e4-92a0-4c42-adc1-77d4599e994b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.174765] env[62569]: DEBUG oslo_vmware.api [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 998.174765] env[62569]: value = "task-1250566" [ 998.174765] env[62569]: _type = "Task" [ 998.174765] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.186176] env[62569]: DEBUG oslo_vmware.api [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250566, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.187227] env[62569]: INFO nova.scheduler.client.report [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Deleted allocations for instance 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e [ 998.199226] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250564, 'name': ReconfigVM_Task, 'duration_secs': 0.662819} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.199568] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 4df2078e-d7ce-4a3f-904b-f3303071ffff/4df2078e-d7ce-4a3f-904b-f3303071ffff.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 998.200233] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f4c43a47-bb99-4e2f-a6d0-6ac5c4e2ce49 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.205846] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 998.205846] env[62569]: value = "task-1250567" [ 998.205846] env[62569]: _type = "Task" [ 998.205846] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.215782] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250567, 'name': Rename_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.339533] env[62569]: DEBUG nova.compute.manager [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 998.364519] env[62569]: DEBUG nova.virt.hardware [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 998.364795] env[62569]: DEBUG nova.virt.hardware [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 998.364962] env[62569]: DEBUG nova.virt.hardware [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 998.365179] env[62569]: DEBUG nova.virt.hardware [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 998.365334] env[62569]: DEBUG nova.virt.hardware [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 998.365488] env[62569]: DEBUG nova.virt.hardware [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 998.365702] env[62569]: DEBUG nova.virt.hardware [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 998.365871] env[62569]: DEBUG nova.virt.hardware [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 998.366063] env[62569]: DEBUG nova.virt.hardware [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 998.366242] env[62569]: DEBUG nova.virt.hardware [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 998.366423] env[62569]: DEBUG nova.virt.hardware [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 998.367592] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cec0084-2019-4200-94a9-1625dee76061 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.375441] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ec453f-30d2-42c3-b159-425c8d1f2d25 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.671264] env[62569]: INFO nova.compute.claims [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 998.687371] env[62569]: DEBUG oslo_vmware.api [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250566, 'name': ReconfigVM_Task, 'duration_secs': 0.153257} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.687707] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 998.687957] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-296fe055-b068-4535-8ed3-aa001ad73bf3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.699421] env[62569]: DEBUG oslo_vmware.api [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 998.699421] env[62569]: value = "task-1250568" [ 998.699421] env[62569]: _type = "Task" [ 998.699421] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.700017] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dfa1248b-211f-4b83-8514-b269d62a5e31 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "93eb1a09-269a-44c8-8f9d-3a2b7e0d715e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.046s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.709945] env[62569]: DEBUG oslo_vmware.api [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250568, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.718867] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250567, 'name': Rename_Task, 'duration_secs': 0.464866} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.719240] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 998.719537] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-08014b2d-cf2d-47ae-945c-32f2aeddd689 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.726260] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 998.726260] env[62569]: value = "task-1250569" [ 998.726260] env[62569]: _type = "Task" [ 998.726260] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.736365] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250569, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.044570] env[62569]: DEBUG nova.compute.manager [req-87a0f17b-2f8b-43b6-9471-e67bc3c9ec5b req-ae147097-863d-4da8-a63a-725c97d8f01c service nova] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Received event network-vif-plugged-8b78baca-9288-4ab0-b21d-7fb96baeaf42 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 999.044797] env[62569]: DEBUG oslo_concurrency.lockutils [req-87a0f17b-2f8b-43b6-9471-e67bc3c9ec5b req-ae147097-863d-4da8-a63a-725c97d8f01c service nova] Acquiring lock "659cc0b8-0f8e-4aa3-83ca-d8c04f355d81-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.045014] env[62569]: DEBUG oslo_concurrency.lockutils [req-87a0f17b-2f8b-43b6-9471-e67bc3c9ec5b req-ae147097-863d-4da8-a63a-725c97d8f01c service nova] Lock "659cc0b8-0f8e-4aa3-83ca-d8c04f355d81-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.045197] env[62569]: DEBUG oslo_concurrency.lockutils [req-87a0f17b-2f8b-43b6-9471-e67bc3c9ec5b req-ae147097-863d-4da8-a63a-725c97d8f01c service nova] Lock "659cc0b8-0f8e-4aa3-83ca-d8c04f355d81-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.045392] env[62569]: DEBUG nova.compute.manager [req-87a0f17b-2f8b-43b6-9471-e67bc3c9ec5b req-ae147097-863d-4da8-a63a-725c97d8f01c service nova] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] No waiting events found dispatching network-vif-plugged-8b78baca-9288-4ab0-b21d-7fb96baeaf42 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 999.045527] env[62569]: WARNING nova.compute.manager [req-87a0f17b-2f8b-43b6-9471-e67bc3c9ec5b req-ae147097-863d-4da8-a63a-725c97d8f01c service nova] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Received unexpected event network-vif-plugged-8b78baca-9288-4ab0-b21d-7fb96baeaf42 for instance with vm_state building and task_state spawning. [ 999.105306] env[62569]: DEBUG oslo_concurrency.lockutils [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.105578] env[62569]: DEBUG oslo_concurrency.lockutils [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.105792] env[62569]: DEBUG oslo_concurrency.lockutils [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "98089a34-074b-4bdb-92ae-f9e23d2551f1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.106017] env[62569]: DEBUG oslo_concurrency.lockutils [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "98089a34-074b-4bdb-92ae-f9e23d2551f1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.106229] env[62569]: DEBUG oslo_concurrency.lockutils [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "98089a34-074b-4bdb-92ae-f9e23d2551f1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.109392] env[62569]: INFO nova.compute.manager [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Terminating instance [ 999.137657] env[62569]: DEBUG nova.network.neutron [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Successfully updated port: 8b78baca-9288-4ab0-b21d-7fb96baeaf42 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 999.179540] env[62569]: INFO nova.compute.resource_tracker [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Updating resource usage from migration a7b12128-5e18-4c5a-a47e-72831dd1bbce [ 999.217147] env[62569]: DEBUG oslo_vmware.api [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250568, 'name': PowerOnVM_Task, 'duration_secs': 0.365898} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.219885] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 999.223392] env[62569]: DEBUG nova.compute.manager [None req-4bbb507b-230b-4904-a635-a64943f8c47f tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 999.224535] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed78d46d-dc66-4eff-8e69-6654a3b796fc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.241494] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250569, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.342298] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ff48a7-9c8f-4570-a0da-a369cfe12dc4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.349954] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-052e3711-a025-461c-9f3c-5e857bc681c2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.380921] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d08585-2842-42f7-b506-ddd27c68a7e9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.387813] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-781005d9-01e6-4d7a-a0eb-76e98527ed4a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.401219] env[62569]: DEBUG nova.compute.provider_tree [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 999.613281] env[62569]: DEBUG nova.compute.manager [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 999.613460] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 999.614396] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-791ddade-4201-4c43-a75b-552965561ab0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.622589] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 999.622879] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6921c10c-0621-4304-9dd6-ca659d29fbfb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.628553] env[62569]: DEBUG oslo_vmware.api [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 999.628553] env[62569]: value = "task-1250570" [ 999.628553] env[62569]: _type = "Task" [ 999.628553] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.637279] env[62569]: DEBUG oslo_vmware.api [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250570, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.639913] env[62569]: DEBUG oslo_concurrency.lockutils [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "refresh_cache-659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.639955] env[62569]: DEBUG oslo_concurrency.lockutils [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired lock "refresh_cache-659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.640123] env[62569]: DEBUG nova.network.neutron [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 999.713661] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "188cac03-4034-4a02-973a-fb1906399fd1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.714068] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "188cac03-4034-4a02-973a-fb1906399fd1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.714304] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "188cac03-4034-4a02-973a-fb1906399fd1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.714515] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "188cac03-4034-4a02-973a-fb1906399fd1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.714736] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "188cac03-4034-4a02-973a-fb1906399fd1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.717030] env[62569]: INFO nova.compute.manager [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Terminating instance [ 999.741021] env[62569]: DEBUG oslo_vmware.api [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250569, 'name': PowerOnVM_Task, 'duration_secs': 0.836786} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.741320] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 999.741566] env[62569]: INFO nova.compute.manager [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Took 8.20 seconds to spawn the instance on the hypervisor. [ 999.741738] env[62569]: DEBUG nova.compute.manager [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 999.742835] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b630c4f7-a923-454e-87d4-21ef8352f5bc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.904736] env[62569]: DEBUG nova.scheduler.client.report [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1000.137764] env[62569]: DEBUG oslo_vmware.api [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250570, 'name': PowerOffVM_Task, 'duration_secs': 0.179338} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.138492] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1000.138492] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1000.138492] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-18d304e5-0b64-4170-bb17-401005fa1449 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.183141] env[62569]: DEBUG nova.network.neutron [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1000.206623] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1000.206867] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1000.207073] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Deleting the datastore file [datastore2] 98089a34-074b-4bdb-92ae-f9e23d2551f1 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1000.209434] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b45245e7-8c94-42b0-9f21-9267cf6b2e6b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.216175] env[62569]: DEBUG oslo_vmware.api [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for the task: (returnval){ [ 1000.216175] env[62569]: value = "task-1250572" [ 1000.216175] env[62569]: _type = "Task" [ 1000.216175] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.224246] env[62569]: DEBUG nova.compute.manager [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1000.224450] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1000.224716] env[62569]: DEBUG oslo_vmware.api [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250572, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.225425] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ec375ec-87ac-42a5-a613-67cd0bb67805 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.231257] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1000.231511] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-30d9750e-0987-4914-b1e8-18dc1ae1d1f5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.236404] env[62569]: DEBUG oslo_vmware.api [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 1000.236404] env[62569]: value = "task-1250573" [ 1000.236404] env[62569]: _type = "Task" [ 1000.236404] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.243635] env[62569]: DEBUG oslo_vmware.api [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250573, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.263101] env[62569]: INFO nova.compute.manager [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Took 18.08 seconds to build instance. [ 1000.362590] env[62569]: DEBUG nova.network.neutron [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Updating instance_info_cache with network_info: [{"id": "8b78baca-9288-4ab0-b21d-7fb96baeaf42", "address": "fa:16:3e:7f:fc:2d", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b78baca-92", "ovs_interfaceid": "8b78baca-9288-4ab0-b21d-7fb96baeaf42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.410034] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.254s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.410912] env[62569]: INFO nova.compute.manager [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Migrating [ 1000.418213] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.733s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.421302] env[62569]: INFO nova.compute.claims [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1000.726663] env[62569]: DEBUG oslo_vmware.api [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Task: {'id': task-1250572, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.175842} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.726949] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1000.727157] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1000.727343] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1000.727526] env[62569]: INFO nova.compute.manager [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1000.727772] env[62569]: DEBUG oslo.service.loopingcall [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1000.727971] env[62569]: DEBUG nova.compute.manager [-] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1000.728082] env[62569]: DEBUG nova.network.neutron [-] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1000.745630] env[62569]: DEBUG oslo_vmware.api [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250573, 'name': PowerOffVM_Task, 'duration_secs': 0.195338} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.745905] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1000.746096] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1000.746349] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d7519d05-8b73-4a4e-89b2-43bb51753c22 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.765530] env[62569]: DEBUG oslo_concurrency.lockutils [None req-850ab616-3b88-4106-8eee-d7c5c9569070 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "4df2078e-d7ce-4a3f-904b-f3303071ffff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.591s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.814450] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1000.814695] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1000.814890] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Deleting the datastore file [datastore2] 188cac03-4034-4a02-973a-fb1906399fd1 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1000.815189] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eea7d17a-ddd8-49e8-8f5c-f34ae50b28a8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.820981] env[62569]: DEBUG oslo_vmware.api [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for the task: (returnval){ [ 1000.820981] env[62569]: value = "task-1250575" [ 1000.820981] env[62569]: _type = "Task" [ 1000.820981] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.829103] env[62569]: DEBUG oslo_vmware.api [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250575, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.865881] env[62569]: DEBUG oslo_concurrency.lockutils [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Releasing lock "refresh_cache-659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.866251] env[62569]: DEBUG nova.compute.manager [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Instance network_info: |[{"id": "8b78baca-9288-4ab0-b21d-7fb96baeaf42", "address": "fa:16:3e:7f:fc:2d", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b78baca-92", "ovs_interfaceid": "8b78baca-9288-4ab0-b21d-7fb96baeaf42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1000.866705] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7f:fc:2d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b356db78-99c7-4464-822c-fc7e193f7878', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8b78baca-9288-4ab0-b21d-7fb96baeaf42', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1000.875433] env[62569]: DEBUG oslo.service.loopingcall [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1000.875663] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1000.875897] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0ceae007-6dfa-4934-98ad-5e1aaebcfa1d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.896829] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1000.896829] env[62569]: value = "task-1250576" [ 1000.896829] env[62569]: _type = "Task" [ 1000.896829] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.904429] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250576, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.934832] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.935049] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.935241] env[62569]: DEBUG nova.network.neutron [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1001.001547] env[62569]: DEBUG oslo_concurrency.lockutils [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquiring lock "0be076c2-176f-47a9-9489-2f7dfe6d1ce3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.001940] env[62569]: DEBUG oslo_concurrency.lockutils [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Lock "0be076c2-176f-47a9-9489-2f7dfe6d1ce3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.193716] env[62569]: DEBUG nova.compute.manager [req-0e057fab-c075-41f9-9e21-2d99564cd5d9 req-e7955edd-dcab-431b-a769-f77805542e1e service nova] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Received event network-changed-8b78baca-9288-4ab0-b21d-7fb96baeaf42 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1001.194023] env[62569]: DEBUG nova.compute.manager [req-0e057fab-c075-41f9-9e21-2d99564cd5d9 req-e7955edd-dcab-431b-a769-f77805542e1e service nova] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Refreshing instance network info cache due to event network-changed-8b78baca-9288-4ab0-b21d-7fb96baeaf42. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1001.194276] env[62569]: DEBUG oslo_concurrency.lockutils [req-0e057fab-c075-41f9-9e21-2d99564cd5d9 req-e7955edd-dcab-431b-a769-f77805542e1e service nova] Acquiring lock "refresh_cache-659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.194425] env[62569]: DEBUG oslo_concurrency.lockutils [req-0e057fab-c075-41f9-9e21-2d99564cd5d9 req-e7955edd-dcab-431b-a769-f77805542e1e service nova] Acquired lock "refresh_cache-659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.194645] env[62569]: DEBUG nova.network.neutron [req-0e057fab-c075-41f9-9e21-2d99564cd5d9 req-e7955edd-dcab-431b-a769-f77805542e1e service nova] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Refreshing network info cache for port 8b78baca-9288-4ab0-b21d-7fb96baeaf42 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1001.331320] env[62569]: DEBUG oslo_vmware.api [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Task: {'id': task-1250575, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.192402} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.331752] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1001.332042] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1001.332325] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1001.332570] env[62569]: INFO nova.compute.manager [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1001.332898] env[62569]: DEBUG oslo.service.loopingcall [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.333134] env[62569]: DEBUG nova.compute.manager [-] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1001.333867] env[62569]: DEBUG nova.network.neutron [-] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1001.409939] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250576, 'name': CreateVM_Task, 'duration_secs': 0.334722} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.410140] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1001.410808] env[62569]: DEBUG oslo_concurrency.lockutils [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.410982] env[62569]: DEBUG oslo_concurrency.lockutils [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.411333] env[62569]: DEBUG oslo_concurrency.lockutils [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1001.411621] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8d944ea-0f96-45bb-bd2a-27dd9de0e1fa {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.418209] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1001.418209] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52fe1b5d-dda1-28d3-f69d-5012cde210ae" [ 1001.418209] env[62569]: _type = "Task" [ 1001.418209] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.427549] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52fe1b5d-dda1-28d3-f69d-5012cde210ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.505981] env[62569]: DEBUG nova.compute.manager [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1001.660866] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-920a4280-5ee4-4021-98b8-1fd52446cbc5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.669042] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84f419e2-524b-47e8-82bf-178f3457df70 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.713106] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baeb886c-0984-4a98-98fb-f88a085d93a4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.723125] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cd2a603-e42c-45bf-96b7-5a6558f34a05 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.740844] env[62569]: DEBUG nova.compute.provider_tree [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1001.744919] env[62569]: DEBUG nova.compute.manager [req-6c97693e-d48d-43db-9b8d-499bb137e3a6 req-478e9d14-0971-433a-9972-36d54ef4dc8b service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Received event network-vif-deleted-91986023-125e-442b-b105-102630ec6036 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1001.746027] env[62569]: INFO nova.compute.manager [req-6c97693e-d48d-43db-9b8d-499bb137e3a6 req-478e9d14-0971-433a-9972-36d54ef4dc8b service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Neutron deleted interface 91986023-125e-442b-b105-102630ec6036; detaching it from the instance and deleting it from the info cache [ 1001.746027] env[62569]: DEBUG nova.network.neutron [req-6c97693e-d48d-43db-9b8d-499bb137e3a6 req-478e9d14-0971-433a-9972-36d54ef4dc8b service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.794297] env[62569]: DEBUG nova.network.neutron [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Updating instance_info_cache with network_info: [{"id": "39e71a4f-3b81-4b34-8cc2-2a5352eca4ce", "address": "fa:16:3e:4b:e2:a8", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap39e71a4f-3b", "ovs_interfaceid": "39e71a4f-3b81-4b34-8cc2-2a5352eca4ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.928358] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52fe1b5d-dda1-28d3-f69d-5012cde210ae, 'name': SearchDatastore_Task, 'duration_secs': 0.010594} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.928665] env[62569]: DEBUG oslo_concurrency.lockutils [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.928906] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1001.929574] env[62569]: DEBUG oslo_concurrency.lockutils [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.929574] env[62569]: DEBUG oslo_concurrency.lockutils [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.929574] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1001.929765] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0ef79661-2c6e-4441-a552-8b799e798d0a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.940347] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1001.940607] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1001.941468] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffa43384-ab70-40d5-b529-cfaf58caec8a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.948228] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1001.948228] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52dea12e-e6eb-0d8f-3bd0-4d99cabf2848" [ 1001.948228] env[62569]: _type = "Task" [ 1001.948228] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.959698] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52dea12e-e6eb-0d8f-3bd0-4d99cabf2848, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.008840] env[62569]: DEBUG nova.network.neutron [req-0e057fab-c075-41f9-9e21-2d99564cd5d9 req-e7955edd-dcab-431b-a769-f77805542e1e service nova] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Updated VIF entry in instance network info cache for port 8b78baca-9288-4ab0-b21d-7fb96baeaf42. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1002.009245] env[62569]: DEBUG nova.network.neutron [req-0e057fab-c075-41f9-9e21-2d99564cd5d9 req-e7955edd-dcab-431b-a769-f77805542e1e service nova] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Updating instance_info_cache with network_info: [{"id": "8b78baca-9288-4ab0-b21d-7fb96baeaf42", "address": "fa:16:3e:7f:fc:2d", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b78baca-92", "ovs_interfaceid": "8b78baca-9288-4ab0-b21d-7fb96baeaf42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.010546] env[62569]: DEBUG nova.network.neutron [-] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.030359] env[62569]: DEBUG oslo_concurrency.lockutils [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.216645] env[62569]: DEBUG nova.network.neutron [-] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.248999] env[62569]: DEBUG nova.scheduler.client.report [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1002.252651] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8c2e97c5-3cc8-4da4-990e-7f0f911b15b4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.264268] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e8c591-736a-4422-b8ec-f23b20b5bf7f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.293403] env[62569]: DEBUG nova.compute.manager [req-6c97693e-d48d-43db-9b8d-499bb137e3a6 req-478e9d14-0971-433a-9972-36d54ef4dc8b service nova] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Detach interface failed, port_id=91986023-125e-442b-b105-102630ec6036, reason: Instance 188cac03-4034-4a02-973a-fb1906399fd1 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1002.297322] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.459321] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "4df2078e-d7ce-4a3f-904b-f3303071ffff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.459603] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "4df2078e-d7ce-4a3f-904b-f3303071ffff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.459807] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "4df2078e-d7ce-4a3f-904b-f3303071ffff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.459999] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "4df2078e-d7ce-4a3f-904b-f3303071ffff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.460195] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "4df2078e-d7ce-4a3f-904b-f3303071ffff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.461905] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52dea12e-e6eb-0d8f-3bd0-4d99cabf2848, 'name': SearchDatastore_Task, 'duration_secs': 0.013372} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.462379] env[62569]: INFO nova.compute.manager [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Terminating instance [ 1002.464990] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-864ad4e5-1e94-4045-8d57-ccfc5dda5fca {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.472091] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1002.472091] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529af35e-3657-c95c-36fb-b385ad53cbba" [ 1002.472091] env[62569]: _type = "Task" [ 1002.472091] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.480591] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529af35e-3657-c95c-36fb-b385ad53cbba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.516666] env[62569]: DEBUG oslo_concurrency.lockutils [req-0e057fab-c075-41f9-9e21-2d99564cd5d9 req-e7955edd-dcab-431b-a769-f77805542e1e service nova] Releasing lock "refresh_cache-659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.517179] env[62569]: INFO nova.compute.manager [-] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Took 1.79 seconds to deallocate network for instance. [ 1002.718818] env[62569]: INFO nova.compute.manager [-] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Took 1.39 seconds to deallocate network for instance. [ 1002.753703] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.335s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.754390] env[62569]: DEBUG nova.compute.manager [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1002.759024] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.566s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.761070] env[62569]: INFO nova.compute.claims [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1002.970276] env[62569]: DEBUG nova.compute.manager [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1002.970382] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1002.971277] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d12492f-f068-4d74-8d64-7da1aa07e350 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.983047] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529af35e-3657-c95c-36fb-b385ad53cbba, 'name': SearchDatastore_Task, 'duration_secs': 0.043546} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.985270] env[62569]: DEBUG oslo_concurrency.lockutils [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.985556] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81/659cc0b8-0f8e-4aa3-83ca-d8c04f355d81.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1002.985863] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1002.986133] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-43e47c11-ceb1-40f7-8ff0-75e5609635c3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.987932] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-71febaac-ce7a-476a-82db-9295e60d0cb5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.994586] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1002.994586] env[62569]: value = "task-1250577" [ 1002.994586] env[62569]: _type = "Task" [ 1002.994586] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.995952] env[62569]: DEBUG oslo_vmware.api [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 1002.995952] env[62569]: value = "task-1250578" [ 1002.995952] env[62569]: _type = "Task" [ 1002.995952] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.006016] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250577, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.008928] env[62569]: DEBUG oslo_vmware.api [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250578, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.023020] env[62569]: DEBUG oslo_concurrency.lockutils [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.222269] env[62569]: DEBUG nova.compute.manager [req-62c02c7e-b945-45a9-963b-ec34d75c5304 req-53581b33-5918-4670-86bc-c1c324607b50 service nova] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Received event network-vif-deleted-f9665162-0615-45cd-b3c2-32be38c9af60 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1003.225300] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.265991] env[62569]: DEBUG nova.compute.utils [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1003.269820] env[62569]: DEBUG nova.compute.manager [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1003.269820] env[62569]: DEBUG nova.network.neutron [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1003.370682] env[62569]: DEBUG nova.policy [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6dd052494f244fd78050c5d26f497720', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2c6a4814c234d53bbd4844cd482fbae', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 1003.510100] env[62569]: DEBUG oslo_vmware.api [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250578, 'name': PowerOffVM_Task, 'duration_secs': 0.177283} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.514248] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1003.514248] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1003.514248] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250577, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.514248] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e3566214-4ca1-4ba8-b952-af9c2b772df8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.580801] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1003.581035] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1003.581266] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Deleting the datastore file [datastore1] 4df2078e-d7ce-4a3f-904b-f3303071ffff {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1003.581535] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f138bba4-dbb1-46e3-a8f9-3f3e9d5e1266 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.589950] env[62569]: DEBUG oslo_vmware.api [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for the task: (returnval){ [ 1003.589950] env[62569]: value = "task-1250580" [ 1003.589950] env[62569]: _type = "Task" [ 1003.589950] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.598178] env[62569]: DEBUG oslo_vmware.api [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250580, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.753703] env[62569]: DEBUG nova.network.neutron [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Successfully created port: 3555d6f1-9762-4670-9cf9-d5d540e928a2 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1003.771583] env[62569]: DEBUG nova.compute.manager [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1003.817108] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8335453-e718-45fb-826b-6b1dd270c6a7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.840725] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Updating instance 'e4eadcdf-a04a-4255-ba1c-fe20156c655f' progress to 0 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1003.963356] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56b50b62-3791-45b6-9f15-2b99438febb8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.970550] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21c430b4-6c91-4e10-b565-6fb7fdca2fdb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.005877] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45920936-53bb-4327-840c-b9ac507f5509 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.015912] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250577, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.77672} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.017301] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3eab964-461c-4800-b597-84a063fc5b2b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.021045] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81/659cc0b8-0f8e-4aa3-83ca-d8c04f355d81.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1004.021283] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1004.021522] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5b484e1c-af83-4cc1-89da-88cf0dbed347 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.032405] env[62569]: DEBUG nova.compute.provider_tree [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.036613] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1004.036613] env[62569]: value = "task-1250581" [ 1004.036613] env[62569]: _type = "Task" [ 1004.036613] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.041823] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250581, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.100681] env[62569]: DEBUG oslo_vmware.api [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Task: {'id': task-1250580, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.503168} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.100969] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1004.101174] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1004.101351] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1004.101534] env[62569]: INFO nova.compute.manager [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1004.101803] env[62569]: DEBUG oslo.service.loopingcall [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.102295] env[62569]: DEBUG nova.compute.manager [-] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1004.102295] env[62569]: DEBUG nova.network.neutron [-] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1004.351019] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1004.351019] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4135aec7-4ea5-4f6d-9f6a-fc3e7a37f864 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.358831] env[62569]: DEBUG oslo_vmware.api [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1004.358831] env[62569]: value = "task-1250582" [ 1004.358831] env[62569]: _type = "Task" [ 1004.358831] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.366029] env[62569]: DEBUG nova.compute.manager [req-3ad926f0-4a46-46d8-a693-ea241c504c7f req-76503cf0-b99b-4adb-8b28-27a37cc43464 service nova] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Received event network-vif-deleted-cc6706bd-0644-4c6c-a181-1b01d3b5f98a {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1004.366029] env[62569]: INFO nova.compute.manager [req-3ad926f0-4a46-46d8-a693-ea241c504c7f req-76503cf0-b99b-4adb-8b28-27a37cc43464 service nova] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Neutron deleted interface cc6706bd-0644-4c6c-a181-1b01d3b5f98a; detaching it from the instance and deleting it from the info cache [ 1004.366029] env[62569]: DEBUG nova.network.neutron [req-3ad926f0-4a46-46d8-a693-ea241c504c7f req-76503cf0-b99b-4adb-8b28-27a37cc43464 service nova] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.372049] env[62569]: DEBUG oslo_vmware.api [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250582, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.536943] env[62569]: DEBUG nova.scheduler.client.report [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1004.549600] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250581, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061105} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.549921] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1004.550735] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e7007eb-8f47-4e29-93d2-d91422e8ab7f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.573945] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81/659cc0b8-0f8e-4aa3-83ca-d8c04f355d81.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1004.574325] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-173b04a6-249a-48dc-a60f-15dcde6939b6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.594860] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1004.594860] env[62569]: value = "task-1250583" [ 1004.594860] env[62569]: _type = "Task" [ 1004.594860] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.603253] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250583, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.783899] env[62569]: DEBUG nova.compute.manager [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1004.811226] env[62569]: DEBUG nova.virt.hardware [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1004.811492] env[62569]: DEBUG nova.virt.hardware [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1004.811731] env[62569]: DEBUG nova.virt.hardware [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1004.811935] env[62569]: DEBUG nova.virt.hardware [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1004.812104] env[62569]: DEBUG nova.virt.hardware [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1004.812258] env[62569]: DEBUG nova.virt.hardware [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1004.812468] env[62569]: DEBUG nova.virt.hardware [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1004.812661] env[62569]: DEBUG nova.virt.hardware [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1004.812839] env[62569]: DEBUG nova.virt.hardware [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1004.813012] env[62569]: DEBUG nova.virt.hardware [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1004.813195] env[62569]: DEBUG nova.virt.hardware [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1004.814077] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fff70f2-e304-472b-aab1-70e884d9a33d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.823164] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75805089-3b9a-4a2f-ba91-0e73d77bb28a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.840739] env[62569]: DEBUG nova.network.neutron [-] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.871273] env[62569]: DEBUG oslo_vmware.api [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250582, 'name': PowerOffVM_Task, 'duration_secs': 0.24785} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.871506] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-526febe6-17a7-4ed1-940a-b8ec3bb58163 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.873290] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1004.873481] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Updating instance 'e4eadcdf-a04a-4255-ba1c-fe20156c655f' progress to 17 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1004.882738] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8baf5a8-c095-4e5d-b018-783059ac0428 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.912858] env[62569]: DEBUG nova.compute.manager [req-3ad926f0-4a46-46d8-a693-ea241c504c7f req-76503cf0-b99b-4adb-8b28-27a37cc43464 service nova] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Detach interface failed, port_id=cc6706bd-0644-4c6c-a181-1b01d3b5f98a, reason: Instance 4df2078e-d7ce-4a3f-904b-f3303071ffff could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1005.042644] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.284s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.043333] env[62569]: DEBUG nova.compute.manager [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1005.046339] env[62569]: DEBUG oslo_concurrency.lockutils [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.016s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.048143] env[62569]: INFO nova.compute.claims [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1005.105237] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250583, 'name': ReconfigVM_Task, 'duration_secs': 0.306849} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.105515] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81/659cc0b8-0f8e-4aa3-83ca-d8c04f355d81.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1005.106181] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-579114b6-8f2a-47de-8916-8066caca603c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.112863] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1005.112863] env[62569]: value = "task-1250584" [ 1005.112863] env[62569]: _type = "Task" [ 1005.112863] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.120931] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250584, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.280777] env[62569]: DEBUG nova.compute.manager [req-5bb0c9c4-95dd-4925-9ae8-16ce563a6742 req-6c81761a-bc41-4b0c-ae9e-7c50d82ee288 service nova] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Received event network-vif-plugged-3555d6f1-9762-4670-9cf9-d5d540e928a2 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1005.280944] env[62569]: DEBUG oslo_concurrency.lockutils [req-5bb0c9c4-95dd-4925-9ae8-16ce563a6742 req-6c81761a-bc41-4b0c-ae9e-7c50d82ee288 service nova] Acquiring lock "4bf2f0fb-420f-4c13-97c1-974af7c49658-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.281173] env[62569]: DEBUG oslo_concurrency.lockutils [req-5bb0c9c4-95dd-4925-9ae8-16ce563a6742 req-6c81761a-bc41-4b0c-ae9e-7c50d82ee288 service nova] Lock "4bf2f0fb-420f-4c13-97c1-974af7c49658-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.281346] env[62569]: DEBUG oslo_concurrency.lockutils [req-5bb0c9c4-95dd-4925-9ae8-16ce563a6742 req-6c81761a-bc41-4b0c-ae9e-7c50d82ee288 service nova] Lock "4bf2f0fb-420f-4c13-97c1-974af7c49658-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.281516] env[62569]: DEBUG nova.compute.manager [req-5bb0c9c4-95dd-4925-9ae8-16ce563a6742 req-6c81761a-bc41-4b0c-ae9e-7c50d82ee288 service nova] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] No waiting events found dispatching network-vif-plugged-3555d6f1-9762-4670-9cf9-d5d540e928a2 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1005.281750] env[62569]: WARNING nova.compute.manager [req-5bb0c9c4-95dd-4925-9ae8-16ce563a6742 req-6c81761a-bc41-4b0c-ae9e-7c50d82ee288 service nova] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Received unexpected event network-vif-plugged-3555d6f1-9762-4670-9cf9-d5d540e928a2 for instance with vm_state building and task_state spawning. [ 1005.343852] env[62569]: INFO nova.compute.manager [-] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Took 1.24 seconds to deallocate network for instance. [ 1005.380344] env[62569]: DEBUG nova.virt.hardware [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1005.380611] env[62569]: DEBUG nova.virt.hardware [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1005.380779] env[62569]: DEBUG nova.virt.hardware [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1005.380966] env[62569]: DEBUG nova.virt.hardware [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1005.384193] env[62569]: DEBUG nova.virt.hardware [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1005.384193] env[62569]: DEBUG nova.virt.hardware [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1005.384193] env[62569]: DEBUG nova.virt.hardware [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1005.384193] env[62569]: DEBUG nova.virt.hardware [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1005.384193] env[62569]: DEBUG nova.virt.hardware [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1005.384193] env[62569]: DEBUG nova.virt.hardware [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1005.384193] env[62569]: DEBUG nova.virt.hardware [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1005.388789] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0271da8-7134-44db-8c42-b880fda99a60 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.410966] env[62569]: DEBUG oslo_vmware.api [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1005.410966] env[62569]: value = "task-1250585" [ 1005.410966] env[62569]: _type = "Task" [ 1005.410966] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.416376] env[62569]: DEBUG nova.network.neutron [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Successfully updated port: 3555d6f1-9762-4670-9cf9-d5d540e928a2 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1005.422964] env[62569]: DEBUG oslo_vmware.api [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250585, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.555689] env[62569]: DEBUG nova.compute.utils [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1005.557178] env[62569]: DEBUG nova.compute.manager [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1005.557352] env[62569]: DEBUG nova.network.neutron [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1005.598250] env[62569]: DEBUG nova.policy [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5da4f6b7a6784a73bd3fed04b275041b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bdba6022e3d4697a336ca28ca4eccec', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 1005.622575] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250584, 'name': Rename_Task, 'duration_secs': 0.152462} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.622841] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1005.623188] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c0ce7e5b-f995-4466-8df4-c75fa47471f6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.629580] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1005.629580] env[62569]: value = "task-1250586" [ 1005.629580] env[62569]: _type = "Task" [ 1005.629580] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.639618] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250586, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.850827] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.927866] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "refresh_cache-4bf2f0fb-420f-4c13-97c1-974af7c49658" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.928099] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired lock "refresh_cache-4bf2f0fb-420f-4c13-97c1-974af7c49658" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.928283] env[62569]: DEBUG nova.network.neutron [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1005.929536] env[62569]: DEBUG oslo_vmware.api [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250585, 'name': ReconfigVM_Task, 'duration_secs': 0.189324} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.929847] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Updating instance 'e4eadcdf-a04a-4255-ba1c-fe20156c655f' progress to 33 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1005.936957] env[62569]: DEBUG nova.network.neutron [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Successfully created port: c2614824-b8a7-4e77-a46e-c9ac6a5dff41 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1006.060758] env[62569]: DEBUG nova.compute.manager [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1006.139210] env[62569]: DEBUG oslo_vmware.api [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250586, 'name': PowerOnVM_Task, 'duration_secs': 0.445268} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.139489] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1006.139691] env[62569]: INFO nova.compute.manager [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Took 7.80 seconds to spawn the instance on the hypervisor. [ 1006.139876] env[62569]: DEBUG nova.compute.manager [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1006.140655] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-987ec76e-355f-4009-85b9-bc48a30b5145 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.253264] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f920f3cf-3748-4727-a41a-e474b4c22e75 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.261687] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28b3f101-6cb1-4660-9b1d-608a19a2a66e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.293070] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8321828-c904-41dc-a8b0-9d20faf3af96 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.300699] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a779b736-2717-4507-81d9-3004b6873f15 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.315596] env[62569]: DEBUG nova.compute.provider_tree [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.439954] env[62569]: DEBUG nova.virt.hardware [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1006.440401] env[62569]: DEBUG nova.virt.hardware [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1006.440709] env[62569]: DEBUG nova.virt.hardware [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1006.441076] env[62569]: DEBUG nova.virt.hardware [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1006.441380] env[62569]: DEBUG nova.virt.hardware [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1006.441700] env[62569]: DEBUG nova.virt.hardware [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1006.442097] env[62569]: DEBUG nova.virt.hardware [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1006.442335] env[62569]: DEBUG nova.virt.hardware [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1006.442695] env[62569]: DEBUG nova.virt.hardware [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1006.443039] env[62569]: DEBUG nova.virt.hardware [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1006.443400] env[62569]: DEBUG nova.virt.hardware [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1006.452758] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Reconfiguring VM instance instance-00000049 to detach disk 2000 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1006.453168] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc69af68-c239-451d-bd5d-36e34cc91488 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.473190] env[62569]: DEBUG oslo_vmware.api [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1006.473190] env[62569]: value = "task-1250587" [ 1006.473190] env[62569]: _type = "Task" [ 1006.473190] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.481940] env[62569]: DEBUG oslo_vmware.api [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250587, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.498964] env[62569]: DEBUG nova.network.neutron [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1006.642175] env[62569]: DEBUG nova.network.neutron [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Updating instance_info_cache with network_info: [{"id": "3555d6f1-9762-4670-9cf9-d5d540e928a2", "address": "fa:16:3e:1f:41:1b", "network": {"id": "884d3248-07f1-4cf7-8ceb-b94b34622384", "bridge": "br-int", "label": "tempest-ServersTestJSON-1882023778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2c6a4814c234d53bbd4844cd482fbae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3555d6f1-97", "ovs_interfaceid": "3555d6f1-9762-4670-9cf9-d5d540e928a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.662705] env[62569]: INFO nova.compute.manager [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Took 14.38 seconds to build instance. [ 1006.819027] env[62569]: DEBUG nova.scheduler.client.report [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1006.983384] env[62569]: DEBUG oslo_vmware.api [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250587, 'name': ReconfigVM_Task, 'duration_secs': 0.171967} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.983841] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Reconfigured VM instance instance-00000049 to detach disk 2000 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1006.984683] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3217321-a9da-4ea3-bdd4-ced5267b8fa0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.013793] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] e4eadcdf-a04a-4255-ba1c-fe20156c655f/e4eadcdf-a04a-4255-ba1c-fe20156c655f.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1007.014407] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-613348a8-55dc-4ebb-8b4a-3f08c595b322 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.033627] env[62569]: DEBUG oslo_vmware.api [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1007.033627] env[62569]: value = "task-1250588" [ 1007.033627] env[62569]: _type = "Task" [ 1007.033627] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.042210] env[62569]: DEBUG oslo_vmware.api [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250588, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.072342] env[62569]: DEBUG nova.compute.manager [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1007.097007] env[62569]: DEBUG nova.virt.hardware [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1007.097276] env[62569]: DEBUG nova.virt.hardware [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1007.097439] env[62569]: DEBUG nova.virt.hardware [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1007.097626] env[62569]: DEBUG nova.virt.hardware [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1007.097778] env[62569]: DEBUG nova.virt.hardware [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1007.097930] env[62569]: DEBUG nova.virt.hardware [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1007.098154] env[62569]: DEBUG nova.virt.hardware [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1007.098319] env[62569]: DEBUG nova.virt.hardware [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1007.098491] env[62569]: DEBUG nova.virt.hardware [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1007.098659] env[62569]: DEBUG nova.virt.hardware [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1007.098837] env[62569]: DEBUG nova.virt.hardware [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1007.100033] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5c2ca09-de89-462f-bc1d-46899cf49ac9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.109689] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f60f6b09-e18c-4565-a1d7-9c818eecec15 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.144889] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Releasing lock "refresh_cache-4bf2f0fb-420f-4c13-97c1-974af7c49658" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.145193] env[62569]: DEBUG nova.compute.manager [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Instance network_info: |[{"id": "3555d6f1-9762-4670-9cf9-d5d540e928a2", "address": "fa:16:3e:1f:41:1b", "network": {"id": "884d3248-07f1-4cf7-8ceb-b94b34622384", "bridge": "br-int", "label": "tempest-ServersTestJSON-1882023778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2c6a4814c234d53bbd4844cd482fbae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3555d6f1-97", "ovs_interfaceid": "3555d6f1-9762-4670-9cf9-d5d540e928a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1007.145576] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1f:41:1b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e9fa4744-8702-4973-b911-ee18192a3e4b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3555d6f1-9762-4670-9cf9-d5d540e928a2', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1007.153122] env[62569]: DEBUG oslo.service.loopingcall [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1007.153325] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1007.153537] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6348169f-3059-4b04-abc7-f2bc6f8f08a4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.167427] env[62569]: DEBUG oslo_concurrency.lockutils [None req-df485552-ce07-4ab4-8e7d-e00da69bde38 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.895s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.173136] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1007.173136] env[62569]: value = "task-1250589" [ 1007.173136] env[62569]: _type = "Task" [ 1007.173136] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.181814] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250589, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.317728] env[62569]: DEBUG nova.compute.manager [req-a34d0b42-da4f-45d7-8187-25e0ed67546d req-8229ab3e-49fb-4352-8fe6-6da4fcf9d6b9 service nova] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Received event network-changed-3555d6f1-9762-4670-9cf9-d5d540e928a2 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1007.317934] env[62569]: DEBUG nova.compute.manager [req-a34d0b42-da4f-45d7-8187-25e0ed67546d req-8229ab3e-49fb-4352-8fe6-6da4fcf9d6b9 service nova] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Refreshing instance network info cache due to event network-changed-3555d6f1-9762-4670-9cf9-d5d540e928a2. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1007.321259] env[62569]: DEBUG oslo_concurrency.lockutils [req-a34d0b42-da4f-45d7-8187-25e0ed67546d req-8229ab3e-49fb-4352-8fe6-6da4fcf9d6b9 service nova] Acquiring lock "refresh_cache-4bf2f0fb-420f-4c13-97c1-974af7c49658" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.321424] env[62569]: DEBUG oslo_concurrency.lockutils [req-a34d0b42-da4f-45d7-8187-25e0ed67546d req-8229ab3e-49fb-4352-8fe6-6da4fcf9d6b9 service nova] Acquired lock "refresh_cache-4bf2f0fb-420f-4c13-97c1-974af7c49658" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.321620] env[62569]: DEBUG nova.network.neutron [req-a34d0b42-da4f-45d7-8187-25e0ed67546d req-8229ab3e-49fb-4352-8fe6-6da4fcf9d6b9 service nova] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Refreshing network info cache for port 3555d6f1-9762-4670-9cf9-d5d540e928a2 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1007.323383] env[62569]: DEBUG oslo_concurrency.lockutils [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.277s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.323841] env[62569]: DEBUG nova.compute.manager [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1007.327266] env[62569]: DEBUG oslo_concurrency.lockutils [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.304s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.327482] env[62569]: DEBUG nova.objects.instance [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lazy-loading 'resources' on Instance uuid 98089a34-074b-4bdb-92ae-f9e23d2551f1 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1007.519679] env[62569]: DEBUG nova.network.neutron [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Successfully updated port: c2614824-b8a7-4e77-a46e-c9ac6a5dff41 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1007.544210] env[62569]: DEBUG oslo_vmware.api [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250588, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.682783] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250589, 'name': CreateVM_Task, 'duration_secs': 0.436241} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.682978] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1007.683741] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.683951] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.684317] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1007.684583] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d791908-c48c-4f54-9e7d-23bb8c7d0546 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.689276] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1007.689276] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]525ef4d7-1476-e01f-20d7-6649fccaf893" [ 1007.689276] env[62569]: _type = "Task" [ 1007.689276] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.697766] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]525ef4d7-1476-e01f-20d7-6649fccaf893, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.831071] env[62569]: DEBUG nova.compute.utils [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1007.835258] env[62569]: DEBUG nova.compute.manager [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1007.835439] env[62569]: DEBUG nova.network.neutron [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1007.879740] env[62569]: DEBUG nova.policy [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a14b0267fd514378ba84cad3d7d39948', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '44464561dec24f209ec4493ac856679b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 1008.025149] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "refresh_cache-c896f38c-f313-491b-86c5-a785721667d7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.025425] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired lock "refresh_cache-c896f38c-f313-491b-86c5-a785721667d7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.026042] env[62569]: DEBUG nova.network.neutron [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1008.043188] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7620d2e-42cd-4514-a703-6161418371b0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.051440] env[62569]: DEBUG oslo_vmware.api [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250588, 'name': ReconfigVM_Task, 'duration_secs': 0.837919} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.053475] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Reconfigured VM instance instance-00000049 to attach disk [datastore1] e4eadcdf-a04a-4255-ba1c-fe20156c655f/e4eadcdf-a04a-4255-ba1c-fe20156c655f.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1008.053840] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Updating instance 'e4eadcdf-a04a-4255-ba1c-fe20156c655f' progress to 50 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1008.057869] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14a396b2-beb5-4af5-9085-e4252d372372 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.091285] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-731079fb-3b74-4c9c-abaa-7150c4331923 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.100842] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52e7f590-73e6-4417-9689-915afcd3926a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.106356] env[62569]: DEBUG nova.network.neutron [req-a34d0b42-da4f-45d7-8187-25e0ed67546d req-8229ab3e-49fb-4352-8fe6-6da4fcf9d6b9 service nova] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Updated VIF entry in instance network info cache for port 3555d6f1-9762-4670-9cf9-d5d540e928a2. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1008.106695] env[62569]: DEBUG nova.network.neutron [req-a34d0b42-da4f-45d7-8187-25e0ed67546d req-8229ab3e-49fb-4352-8fe6-6da4fcf9d6b9 service nova] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Updating instance_info_cache with network_info: [{"id": "3555d6f1-9762-4670-9cf9-d5d540e928a2", "address": "fa:16:3e:1f:41:1b", "network": {"id": "884d3248-07f1-4cf7-8ceb-b94b34622384", "bridge": "br-int", "label": "tempest-ServersTestJSON-1882023778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2c6a4814c234d53bbd4844cd482fbae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3555d6f1-97", "ovs_interfaceid": "3555d6f1-9762-4670-9cf9-d5d540e928a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.119223] env[62569]: DEBUG nova.compute.provider_tree [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1008.153927] env[62569]: DEBUG nova.network.neutron [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Successfully created port: 2e8ed1b5-2a5a-4084-b227-177a876c7cf0 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1008.202948] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]525ef4d7-1476-e01f-20d7-6649fccaf893, 'name': SearchDatastore_Task, 'duration_secs': 0.011837} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.203393] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.203737] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1008.204112] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.204354] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.204630] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1008.204983] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f7382b93-bf76-4340-8ef8-a7dad378ad38 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.214260] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1008.214527] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1008.215568] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20d2beba-f054-4522-b05c-09688e16516b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.222124] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1008.222124] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528ed12c-e7cc-1c2a-1ac8-a96922d5fb54" [ 1008.222124] env[62569]: _type = "Task" [ 1008.222124] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.232554] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528ed12c-e7cc-1c2a-1ac8-a96922d5fb54, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.339075] env[62569]: DEBUG nova.compute.manager [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1008.563861] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c0110eb-85c0-42a0-a7a7-03f25c5f204c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.583100] env[62569]: DEBUG nova.network.neutron [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1008.585410] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b69ad860-ef30-4426-8b73-4da44e3602dc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.603833] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Updating instance 'e4eadcdf-a04a-4255-ba1c-fe20156c655f' progress to 67 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1008.610026] env[62569]: DEBUG oslo_concurrency.lockutils [req-a34d0b42-da4f-45d7-8187-25e0ed67546d req-8229ab3e-49fb-4352-8fe6-6da4fcf9d6b9 service nova] Releasing lock "refresh_cache-4bf2f0fb-420f-4c13-97c1-974af7c49658" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.610026] env[62569]: DEBUG nova.compute.manager [req-a34d0b42-da4f-45d7-8187-25e0ed67546d req-8229ab3e-49fb-4352-8fe6-6da4fcf9d6b9 service nova] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Received event network-changed-8b78baca-9288-4ab0-b21d-7fb96baeaf42 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1008.610026] env[62569]: DEBUG nova.compute.manager [req-a34d0b42-da4f-45d7-8187-25e0ed67546d req-8229ab3e-49fb-4352-8fe6-6da4fcf9d6b9 service nova] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Refreshing instance network info cache due to event network-changed-8b78baca-9288-4ab0-b21d-7fb96baeaf42. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1008.610178] env[62569]: DEBUG oslo_concurrency.lockutils [req-a34d0b42-da4f-45d7-8187-25e0ed67546d req-8229ab3e-49fb-4352-8fe6-6da4fcf9d6b9 service nova] Acquiring lock "refresh_cache-659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.610276] env[62569]: DEBUG oslo_concurrency.lockutils [req-a34d0b42-da4f-45d7-8187-25e0ed67546d req-8229ab3e-49fb-4352-8fe6-6da4fcf9d6b9 service nova] Acquired lock "refresh_cache-659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.610379] env[62569]: DEBUG nova.network.neutron [req-a34d0b42-da4f-45d7-8187-25e0ed67546d req-8229ab3e-49fb-4352-8fe6-6da4fcf9d6b9 service nova] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Refreshing network info cache for port 8b78baca-9288-4ab0-b21d-7fb96baeaf42 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1008.622436] env[62569]: DEBUG nova.scheduler.client.report [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1008.727161] env[62569]: DEBUG nova.network.neutron [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Updating instance_info_cache with network_info: [{"id": "c2614824-b8a7-4e77-a46e-c9ac6a5dff41", "address": "fa:16:3e:6f:d5:b0", "network": {"id": "30d3bf4a-8f55-492f-a049-181e29731818", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1244949475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bdba6022e3d4697a336ca28ca4eccec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2614824-b8", "ovs_interfaceid": "c2614824-b8a7-4e77-a46e-c9ac6a5dff41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.737056] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]528ed12c-e7cc-1c2a-1ac8-a96922d5fb54, 'name': SearchDatastore_Task, 'duration_secs': 0.012675} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.737269] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc36d9ef-c59f-466b-9294-8e0493284ad8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.743227] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1008.743227] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]522fa0fc-3941-f005-6e80-7197eb995f6e" [ 1008.743227] env[62569]: _type = "Task" [ 1008.743227] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.751908] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]522fa0fc-3941-f005-6e80-7197eb995f6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.126355] env[62569]: DEBUG oslo_concurrency.lockutils [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.799s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.128466] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.904s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.128702] env[62569]: DEBUG nova.objects.instance [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lazy-loading 'resources' on Instance uuid 188cac03-4034-4a02-973a-fb1906399fd1 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1009.153029] env[62569]: INFO nova.scheduler.client.report [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Deleted allocations for instance 98089a34-074b-4bdb-92ae-f9e23d2551f1 [ 1009.158961] env[62569]: DEBUG nova.network.neutron [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Port 39e71a4f-3b81-4b34-8cc2-2a5352eca4ce binding to destination host cpu-1 is already ACTIVE {{(pid=62569) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1009.231336] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Releasing lock "refresh_cache-c896f38c-f313-491b-86c5-a785721667d7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.231672] env[62569]: DEBUG nova.compute.manager [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Instance network_info: |[{"id": "c2614824-b8a7-4e77-a46e-c9ac6a5dff41", "address": "fa:16:3e:6f:d5:b0", "network": {"id": "30d3bf4a-8f55-492f-a049-181e29731818", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1244949475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bdba6022e3d4697a336ca28ca4eccec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2614824-b8", "ovs_interfaceid": "c2614824-b8a7-4e77-a46e-c9ac6a5dff41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1009.232125] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6f:d5:b0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '983826cf-6390-4ec6-bf97-30a1060947fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c2614824-b8a7-4e77-a46e-c9ac6a5dff41', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1009.239547] env[62569]: DEBUG oslo.service.loopingcall [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.240205] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c896f38c-f313-491b-86c5-a785721667d7] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1009.240302] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-46c38e89-194c-4fd8-8c6d-994e8808cb4f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.266949] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]522fa0fc-3941-f005-6e80-7197eb995f6e, 'name': SearchDatastore_Task, 'duration_secs': 0.009352} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.268278] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.268562] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 4bf2f0fb-420f-4c13-97c1-974af7c49658/4bf2f0fb-420f-4c13-97c1-974af7c49658.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1009.268803] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1009.268803] env[62569]: value = "task-1250590" [ 1009.268803] env[62569]: _type = "Task" [ 1009.268803] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.268998] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4cf80ee9-3a1f-4e5e-979c-30276a9df87b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.280072] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250590, 'name': CreateVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.281344] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1009.281344] env[62569]: value = "task-1250591" [ 1009.281344] env[62569]: _type = "Task" [ 1009.281344] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.289367] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250591, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.342828] env[62569]: DEBUG nova.compute.manager [req-dbb5f7f2-ced4-4a86-bb2b-4a239286e0de req-fe3efc97-9f70-4bf9-86f9-68d64e5c5fab service nova] [instance: c896f38c-f313-491b-86c5-a785721667d7] Received event network-vif-plugged-c2614824-b8a7-4e77-a46e-c9ac6a5dff41 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1009.342983] env[62569]: DEBUG oslo_concurrency.lockutils [req-dbb5f7f2-ced4-4a86-bb2b-4a239286e0de req-fe3efc97-9f70-4bf9-86f9-68d64e5c5fab service nova] Acquiring lock "c896f38c-f313-491b-86c5-a785721667d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.343225] env[62569]: DEBUG oslo_concurrency.lockutils [req-dbb5f7f2-ced4-4a86-bb2b-4a239286e0de req-fe3efc97-9f70-4bf9-86f9-68d64e5c5fab service nova] Lock "c896f38c-f313-491b-86c5-a785721667d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.343418] env[62569]: DEBUG oslo_concurrency.lockutils [req-dbb5f7f2-ced4-4a86-bb2b-4a239286e0de req-fe3efc97-9f70-4bf9-86f9-68d64e5c5fab service nova] Lock "c896f38c-f313-491b-86c5-a785721667d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.343610] env[62569]: DEBUG nova.compute.manager [req-dbb5f7f2-ced4-4a86-bb2b-4a239286e0de req-fe3efc97-9f70-4bf9-86f9-68d64e5c5fab service nova] [instance: c896f38c-f313-491b-86c5-a785721667d7] No waiting events found dispatching network-vif-plugged-c2614824-b8a7-4e77-a46e-c9ac6a5dff41 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1009.343797] env[62569]: WARNING nova.compute.manager [req-dbb5f7f2-ced4-4a86-bb2b-4a239286e0de req-fe3efc97-9f70-4bf9-86f9-68d64e5c5fab service nova] [instance: c896f38c-f313-491b-86c5-a785721667d7] Received unexpected event network-vif-plugged-c2614824-b8a7-4e77-a46e-c9ac6a5dff41 for instance with vm_state building and task_state spawning. [ 1009.344027] env[62569]: DEBUG nova.compute.manager [req-dbb5f7f2-ced4-4a86-bb2b-4a239286e0de req-fe3efc97-9f70-4bf9-86f9-68d64e5c5fab service nova] [instance: c896f38c-f313-491b-86c5-a785721667d7] Received event network-changed-c2614824-b8a7-4e77-a46e-c9ac6a5dff41 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1009.344209] env[62569]: DEBUG nova.compute.manager [req-dbb5f7f2-ced4-4a86-bb2b-4a239286e0de req-fe3efc97-9f70-4bf9-86f9-68d64e5c5fab service nova] [instance: c896f38c-f313-491b-86c5-a785721667d7] Refreshing instance network info cache due to event network-changed-c2614824-b8a7-4e77-a46e-c9ac6a5dff41. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1009.344418] env[62569]: DEBUG oslo_concurrency.lockutils [req-dbb5f7f2-ced4-4a86-bb2b-4a239286e0de req-fe3efc97-9f70-4bf9-86f9-68d64e5c5fab service nova] Acquiring lock "refresh_cache-c896f38c-f313-491b-86c5-a785721667d7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.344571] env[62569]: DEBUG oslo_concurrency.lockutils [req-dbb5f7f2-ced4-4a86-bb2b-4a239286e0de req-fe3efc97-9f70-4bf9-86f9-68d64e5c5fab service nova] Acquired lock "refresh_cache-c896f38c-f313-491b-86c5-a785721667d7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.344749] env[62569]: DEBUG nova.network.neutron [req-dbb5f7f2-ced4-4a86-bb2b-4a239286e0de req-fe3efc97-9f70-4bf9-86f9-68d64e5c5fab service nova] [instance: c896f38c-f313-491b-86c5-a785721667d7] Refreshing network info cache for port c2614824-b8a7-4e77-a46e-c9ac6a5dff41 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1009.349725] env[62569]: DEBUG nova.compute.manager [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1009.379108] env[62569]: DEBUG nova.virt.hardware [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1009.380104] env[62569]: DEBUG nova.virt.hardware [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1009.380104] env[62569]: DEBUG nova.virt.hardware [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1009.380104] env[62569]: DEBUG nova.virt.hardware [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1009.380295] env[62569]: DEBUG nova.virt.hardware [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1009.384019] env[62569]: DEBUG nova.virt.hardware [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1009.384019] env[62569]: DEBUG nova.virt.hardware [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1009.384019] env[62569]: DEBUG nova.virt.hardware [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1009.384019] env[62569]: DEBUG nova.virt.hardware [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1009.384019] env[62569]: DEBUG nova.virt.hardware [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1009.384019] env[62569]: DEBUG nova.virt.hardware [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1009.384019] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3e8f845-792a-461d-848d-74eb9e180ff0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.392132] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-315f8847-81de-44c4-b3aa-5576653bb8a6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.591683] env[62569]: DEBUG nova.network.neutron [req-a34d0b42-da4f-45d7-8187-25e0ed67546d req-8229ab3e-49fb-4352-8fe6-6da4fcf9d6b9 service nova] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Updated VIF entry in instance network info cache for port 8b78baca-9288-4ab0-b21d-7fb96baeaf42. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1009.592241] env[62569]: DEBUG nova.network.neutron [req-a34d0b42-da4f-45d7-8187-25e0ed67546d req-8229ab3e-49fb-4352-8fe6-6da4fcf9d6b9 service nova] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Updating instance_info_cache with network_info: [{"id": "8b78baca-9288-4ab0-b21d-7fb96baeaf42", "address": "fa:16:3e:7f:fc:2d", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b78baca-92", "ovs_interfaceid": "8b78baca-9288-4ab0-b21d-7fb96baeaf42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.661265] env[62569]: DEBUG oslo_concurrency.lockutils [None req-610f6c68-3b5d-403f-9b3b-de58ca53e2e9 tempest-AttachVolumeTestJSON-2114587466 tempest-AttachVolumeTestJSON-2114587466-project-member] Lock "98089a34-074b-4bdb-92ae-f9e23d2551f1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.555s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.739706] env[62569]: DEBUG nova.compute.manager [req-593d0ebf-dfff-4356-a621-27a1a82b5b78 req-145c2047-6c6d-4854-9941-16a53023b741 service nova] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Received event network-vif-plugged-2e8ed1b5-2a5a-4084-b227-177a876c7cf0 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1009.739706] env[62569]: DEBUG oslo_concurrency.lockutils [req-593d0ebf-dfff-4356-a621-27a1a82b5b78 req-145c2047-6c6d-4854-9941-16a53023b741 service nova] Acquiring lock "0be076c2-176f-47a9-9489-2f7dfe6d1ce3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.739706] env[62569]: DEBUG oslo_concurrency.lockutils [req-593d0ebf-dfff-4356-a621-27a1a82b5b78 req-145c2047-6c6d-4854-9941-16a53023b741 service nova] Lock "0be076c2-176f-47a9-9489-2f7dfe6d1ce3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.739706] env[62569]: DEBUG oslo_concurrency.lockutils [req-593d0ebf-dfff-4356-a621-27a1a82b5b78 req-145c2047-6c6d-4854-9941-16a53023b741 service nova] Lock "0be076c2-176f-47a9-9489-2f7dfe6d1ce3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.739871] env[62569]: DEBUG nova.compute.manager [req-593d0ebf-dfff-4356-a621-27a1a82b5b78 req-145c2047-6c6d-4854-9941-16a53023b741 service nova] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] No waiting events found dispatching network-vif-plugged-2e8ed1b5-2a5a-4084-b227-177a876c7cf0 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1009.740008] env[62569]: WARNING nova.compute.manager [req-593d0ebf-dfff-4356-a621-27a1a82b5b78 req-145c2047-6c6d-4854-9941-16a53023b741 service nova] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Received unexpected event network-vif-plugged-2e8ed1b5-2a5a-4084-b227-177a876c7cf0 for instance with vm_state building and task_state spawning. [ 1009.782988] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250590, 'name': CreateVM_Task, 'duration_secs': 0.317967} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.787014] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c896f38c-f313-491b-86c5-a785721667d7] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1009.788030] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.788030] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.788228] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1009.790906] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31396b97-0482-4761-827e-f74ddaa9276a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.795824] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250591, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.456683} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.796129] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 4bf2f0fb-420f-4c13-97c1-974af7c49658/4bf2f0fb-420f-4c13-97c1-974af7c49658.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1009.796350] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1009.796585] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fc9cabe6-8379-49e1-8a87-ae87971a0fd5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.799555] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1009.799555] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52880dc0-46ff-24cd-7f5c-24d9b8025b36" [ 1009.799555] env[62569]: _type = "Task" [ 1009.799555] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.805831] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1009.805831] env[62569]: value = "task-1250592" [ 1009.805831] env[62569]: _type = "Task" [ 1009.805831] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.817020] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52880dc0-46ff-24cd-7f5c-24d9b8025b36, 'name': SearchDatastore_Task, 'duration_secs': 0.008703} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.817020] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.817020] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1009.817020] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.817020] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.817020] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1009.821668] env[62569]: DEBUG nova.network.neutron [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Successfully updated port: 2e8ed1b5-2a5a-4084-b227-177a876c7cf0 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1009.822461] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-26e1fe95-7f4a-4ac8-ac5e-5e7819a49e60 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.824418] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250592, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.832768] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1009.832884] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1009.833569] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62aec6fa-9fa4-43a0-8eff-219c674a175d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.840711] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1009.840711] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b21db1-46e1-97e2-cd58-19f5864066ff" [ 1009.840711] env[62569]: _type = "Task" [ 1009.840711] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.849276] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b21db1-46e1-97e2-cd58-19f5864066ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.850943] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79555b6b-56d5-46e7-82c8-1ca582877980 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.860621] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-462486f7-d375-4e14-b8ff-082830331db1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.915354] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ba63e51-a0a9-4c53-a4ea-05e86b7568c2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.926525] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-296e6992-70e0-4c87-b5c3-992b4eeec194 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.946630] env[62569]: DEBUG nova.compute.provider_tree [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1010.095604] env[62569]: DEBUG oslo_concurrency.lockutils [req-a34d0b42-da4f-45d7-8187-25e0ed67546d req-8229ab3e-49fb-4352-8fe6-6da4fcf9d6b9 service nova] Releasing lock "refresh_cache-659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.100330] env[62569]: DEBUG nova.network.neutron [req-dbb5f7f2-ced4-4a86-bb2b-4a239286e0de req-fe3efc97-9f70-4bf9-86f9-68d64e5c5fab service nova] [instance: c896f38c-f313-491b-86c5-a785721667d7] Updated VIF entry in instance network info cache for port c2614824-b8a7-4e77-a46e-c9ac6a5dff41. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1010.100680] env[62569]: DEBUG nova.network.neutron [req-dbb5f7f2-ced4-4a86-bb2b-4a239286e0de req-fe3efc97-9f70-4bf9-86f9-68d64e5c5fab service nova] [instance: c896f38c-f313-491b-86c5-a785721667d7] Updating instance_info_cache with network_info: [{"id": "c2614824-b8a7-4e77-a46e-c9ac6a5dff41", "address": "fa:16:3e:6f:d5:b0", "network": {"id": "30d3bf4a-8f55-492f-a049-181e29731818", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1244949475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bdba6022e3d4697a336ca28ca4eccec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2614824-b8", "ovs_interfaceid": "c2614824-b8a7-4e77-a46e-c9ac6a5dff41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.185024] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.185024] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.185024] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.317032] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250592, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061795} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.317373] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1010.318455] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-339a7009-f03d-48aa-9fe2-9e9a597fa44f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.333311] env[62569]: DEBUG oslo_concurrency.lockutils [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquiring lock "refresh_cache-0be076c2-176f-47a9-9489-2f7dfe6d1ce3" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.333311] env[62569]: DEBUG oslo_concurrency.lockutils [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquired lock "refresh_cache-0be076c2-176f-47a9-9489-2f7dfe6d1ce3" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.333311] env[62569]: DEBUG nova.network.neutron [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1010.343935] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 4bf2f0fb-420f-4c13-97c1-974af7c49658/4bf2f0fb-420f-4c13-97c1-974af7c49658.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1010.345110] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4933357c-472a-4f7a-8333-abd1aed3370a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.368711] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b21db1-46e1-97e2-cd58-19f5864066ff, 'name': SearchDatastore_Task, 'duration_secs': 0.008066} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.371066] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1010.371066] env[62569]: value = "task-1250594" [ 1010.371066] env[62569]: _type = "Task" [ 1010.371066] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.371320] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a77ef1a5-77bb-4338-947e-96fd2dc41712 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.380228] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1010.380228] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5292cf95-5662-6043-fee1-3bcca1ccbb9a" [ 1010.380228] env[62569]: _type = "Task" [ 1010.380228] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.383264] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250594, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.391441] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5292cf95-5662-6043-fee1-3bcca1ccbb9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.449963] env[62569]: DEBUG nova.scheduler.client.report [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1010.603559] env[62569]: DEBUG oslo_concurrency.lockutils [req-dbb5f7f2-ced4-4a86-bb2b-4a239286e0de req-fe3efc97-9f70-4bf9-86f9-68d64e5c5fab service nova] Releasing lock "refresh_cache-c896f38c-f313-491b-86c5-a785721667d7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.873699] env[62569]: DEBUG nova.network.neutron [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1010.885025] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250594, 'name': ReconfigVM_Task, 'duration_secs': 0.321008} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.887809] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 4bf2f0fb-420f-4c13-97c1-974af7c49658/4bf2f0fb-420f-4c13-97c1-974af7c49658.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1010.888427] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f32b4a4e-3c60-4a1a-8ad1-572b6ab694ba {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.899074] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5292cf95-5662-6043-fee1-3bcca1ccbb9a, 'name': SearchDatastore_Task, 'duration_secs': 0.010305} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.899389] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1010.899389] env[62569]: value = "task-1250595" [ 1010.899389] env[62569]: _type = "Task" [ 1010.899389] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.899697] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.899985] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] c896f38c-f313-491b-86c5-a785721667d7/c896f38c-f313-491b-86c5-a785721667d7.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1010.900303] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5163a28e-d79f-43e5-827c-65205397c6ea {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.909913] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250595, 'name': Rename_Task} progress is 10%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.911097] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1010.911097] env[62569]: value = "task-1250596" [ 1010.911097] env[62569]: _type = "Task" [ 1010.911097] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.918645] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250596, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.955487] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.827s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.958108] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.107s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.958436] env[62569]: DEBUG nova.objects.instance [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lazy-loading 'resources' on Instance uuid 4df2078e-d7ce-4a3f-904b-f3303071ffff {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1010.986132] env[62569]: INFO nova.scheduler.client.report [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Deleted allocations for instance 188cac03-4034-4a02-973a-fb1906399fd1 [ 1011.014120] env[62569]: DEBUG nova.network.neutron [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Updating instance_info_cache with network_info: [{"id": "2e8ed1b5-2a5a-4084-b227-177a876c7cf0", "address": "fa:16:3e:93:fd:07", "network": {"id": "73711c83-88d7-490c-bc06-44e6ed11339c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1710731049-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "44464561dec24f209ec4493ac856679b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "05b1253d-2b87-4158-9ff1-dafcf829f11f", "external-id": "nsx-vlan-transportzone-55", "segmentation_id": 55, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e8ed1b5-2a", "ovs_interfaceid": "2e8ed1b5-2a5a-4084-b227-177a876c7cf0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.224762] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.224998] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.225238] env[62569]: DEBUG nova.network.neutron [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1011.414016] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250595, 'name': Rename_Task, 'duration_secs': 0.132164} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.414016] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1011.414016] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-994ee9dd-fc7a-46e5-8f1a-c681c76bd27a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.420831] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250596, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.4382} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.422167] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] c896f38c-f313-491b-86c5-a785721667d7/c896f38c-f313-491b-86c5-a785721667d7.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1011.422539] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1011.423171] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1011.423171] env[62569]: value = "task-1250597" [ 1011.423171] env[62569]: _type = "Task" [ 1011.423171] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.423616] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1d76dcc3-d38b-4a94-969c-9acca10da7b3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.439022] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250597, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.439022] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1011.439022] env[62569]: value = "task-1250598" [ 1011.439022] env[62569]: _type = "Task" [ 1011.439022] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.444592] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250598, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.498521] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9acbe25b-de90-4db5-b194-fb1b551b03c3 tempest-AttachInterfacesTestJSON-1486200533 tempest-AttachInterfacesTestJSON-1486200533-project-member] Lock "188cac03-4034-4a02-973a-fb1906399fd1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.784s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.518522] env[62569]: DEBUG oslo_concurrency.lockutils [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Releasing lock "refresh_cache-0be076c2-176f-47a9-9489-2f7dfe6d1ce3" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.518837] env[62569]: DEBUG nova.compute.manager [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Instance network_info: |[{"id": "2e8ed1b5-2a5a-4084-b227-177a876c7cf0", "address": "fa:16:3e:93:fd:07", "network": {"id": "73711c83-88d7-490c-bc06-44e6ed11339c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1710731049-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "44464561dec24f209ec4493ac856679b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "05b1253d-2b87-4158-9ff1-dafcf829f11f", "external-id": "nsx-vlan-transportzone-55", "segmentation_id": 55, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e8ed1b5-2a", "ovs_interfaceid": "2e8ed1b5-2a5a-4084-b227-177a876c7cf0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1011.519500] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:93:fd:07', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '05b1253d-2b87-4158-9ff1-dafcf829f11f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2e8ed1b5-2a5a-4084-b227-177a876c7cf0', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1011.527569] env[62569]: DEBUG oslo.service.loopingcall [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1011.528081] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1011.528724] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4cf456b9-267d-4b20-83da-43511b89a732 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.550701] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1011.550701] env[62569]: value = "task-1250599" [ 1011.550701] env[62569]: _type = "Task" [ 1011.550701] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.559229] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250599, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.643929] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8e09fd8-9975-4fdd-8bae-5cfc307b60d8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.653540] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74869457-f182-470f-97e1-c47783636c50 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.688386] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98d18fcb-4f85-42a0-baee-010e33207f0b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.697031] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d20a89-9895-4561-b245-ba31e26ff01c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.711760] env[62569]: DEBUG nova.compute.provider_tree [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.773587] env[62569]: DEBUG nova.compute.manager [req-364fcc1a-c211-4cf3-9abd-2b40562e76de req-57376ab7-77b9-4276-8214-dd80db92e0a8 service nova] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Received event network-changed-2e8ed1b5-2a5a-4084-b227-177a876c7cf0 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1011.773810] env[62569]: DEBUG nova.compute.manager [req-364fcc1a-c211-4cf3-9abd-2b40562e76de req-57376ab7-77b9-4276-8214-dd80db92e0a8 service nova] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Refreshing instance network info cache due to event network-changed-2e8ed1b5-2a5a-4084-b227-177a876c7cf0. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1011.774058] env[62569]: DEBUG oslo_concurrency.lockutils [req-364fcc1a-c211-4cf3-9abd-2b40562e76de req-57376ab7-77b9-4276-8214-dd80db92e0a8 service nova] Acquiring lock "refresh_cache-0be076c2-176f-47a9-9489-2f7dfe6d1ce3" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.774226] env[62569]: DEBUG oslo_concurrency.lockutils [req-364fcc1a-c211-4cf3-9abd-2b40562e76de req-57376ab7-77b9-4276-8214-dd80db92e0a8 service nova] Acquired lock "refresh_cache-0be076c2-176f-47a9-9489-2f7dfe6d1ce3" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.774839] env[62569]: DEBUG nova.network.neutron [req-364fcc1a-c211-4cf3-9abd-2b40562e76de req-57376ab7-77b9-4276-8214-dd80db92e0a8 service nova] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Refreshing network info cache for port 2e8ed1b5-2a5a-4084-b227-177a876c7cf0 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1011.938417] env[62569]: DEBUG oslo_vmware.api [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250597, 'name': PowerOnVM_Task, 'duration_secs': 0.507648} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.938417] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1011.938417] env[62569]: INFO nova.compute.manager [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Took 7.15 seconds to spawn the instance on the hypervisor. [ 1011.938417] env[62569]: DEBUG nova.compute.manager [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1011.938417] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1e61b6-acd0-452f-a27d-f68db1770dc8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.954078] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250598, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064087} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.954602] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1011.955586] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da2bf9e0-a3e0-4196-9e94-7c9ed33cd6fc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.977801] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] c896f38c-f313-491b-86c5-a785721667d7/c896f38c-f313-491b-86c5-a785721667d7.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1011.980337] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9196413a-9abc-4341-a4ba-6cd8c132abc9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.006789] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1012.006789] env[62569]: value = "task-1250600" [ 1012.006789] env[62569]: _type = "Task" [ 1012.006789] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.018832] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250600, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.050952] env[62569]: DEBUG nova.network.neutron [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Updating instance_info_cache with network_info: [{"id": "39e71a4f-3b81-4b34-8cc2-2a5352eca4ce", "address": "fa:16:3e:4b:e2:a8", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap39e71a4f-3b", "ovs_interfaceid": "39e71a4f-3b81-4b34-8cc2-2a5352eca4ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.062538] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250599, 'name': CreateVM_Task} progress is 25%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.215079] env[62569]: DEBUG nova.scheduler.client.report [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1012.470410] env[62569]: INFO nova.compute.manager [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Took 15.80 seconds to build instance. [ 1012.508145] env[62569]: DEBUG nova.network.neutron [req-364fcc1a-c211-4cf3-9abd-2b40562e76de req-57376ab7-77b9-4276-8214-dd80db92e0a8 service nova] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Updated VIF entry in instance network info cache for port 2e8ed1b5-2a5a-4084-b227-177a876c7cf0. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1012.508145] env[62569]: DEBUG nova.network.neutron [req-364fcc1a-c211-4cf3-9abd-2b40562e76de req-57376ab7-77b9-4276-8214-dd80db92e0a8 service nova] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Updating instance_info_cache with network_info: [{"id": "2e8ed1b5-2a5a-4084-b227-177a876c7cf0", "address": "fa:16:3e:93:fd:07", "network": {"id": "73711c83-88d7-490c-bc06-44e6ed11339c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1710731049-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "44464561dec24f209ec4493ac856679b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "05b1253d-2b87-4158-9ff1-dafcf829f11f", "external-id": "nsx-vlan-transportzone-55", "segmentation_id": 55, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e8ed1b5-2a", "ovs_interfaceid": "2e8ed1b5-2a5a-4084-b227-177a876c7cf0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.520876] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250600, 'name': ReconfigVM_Task, 'duration_secs': 0.267954} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.521723] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Reconfigured VM instance instance-00000068 to attach disk [datastore2] c896f38c-f313-491b-86c5-a785721667d7/c896f38c-f313-491b-86c5-a785721667d7.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1012.522552] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-14a0f1c5-751d-4633-a4d1-19bf171854f9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.530018] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1012.530018] env[62569]: value = "task-1250601" [ 1012.530018] env[62569]: _type = "Task" [ 1012.530018] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.539207] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250601, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.553696] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.568812] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250599, 'name': CreateVM_Task, 'duration_secs': 0.700042} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.570029] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1012.570029] env[62569]: DEBUG oslo_concurrency.lockutils [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.570029] env[62569]: DEBUG oslo_concurrency.lockutils [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.570205] env[62569]: DEBUG oslo_concurrency.lockutils [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1012.571110] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec07f911-e23e-4754-bb15-b4661cc21f81 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.575136] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 1012.575136] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b3dd4e-5974-2526-1f47-d36e134e7ccd" [ 1012.575136] env[62569]: _type = "Task" [ 1012.575136] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.584110] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b3dd4e-5974-2526-1f47-d36e134e7ccd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.722554] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.762s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.744260] env[62569]: INFO nova.scheduler.client.report [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Deleted allocations for instance 4df2078e-d7ce-4a3f-904b-f3303071ffff [ 1012.972706] env[62569]: DEBUG oslo_concurrency.lockutils [None req-2dce9b84-4cc1-4377-9edd-515bd4253e1f tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "4bf2f0fb-420f-4c13-97c1-974af7c49658" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.314s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.012433] env[62569]: DEBUG oslo_concurrency.lockutils [req-364fcc1a-c211-4cf3-9abd-2b40562e76de req-57376ab7-77b9-4276-8214-dd80db92e0a8 service nova] Releasing lock "refresh_cache-0be076c2-176f-47a9-9489-2f7dfe6d1ce3" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.041486] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250601, 'name': Rename_Task, 'duration_secs': 0.13358} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.043822] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1013.043822] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-315789c7-cc10-4060-962a-192131249fc5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.051017] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1013.051017] env[62569]: value = "task-1250602" [ 1013.051017] env[62569]: _type = "Task" [ 1013.051017] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.057665] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250602, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.081757] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61277478-cf60-494a-ba59-2586fddca3eb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.112016] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-159b48f7-b2ff-46c8-8a50-68a053fc1135 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.113964] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b3dd4e-5974-2526-1f47-d36e134e7ccd, 'name': SearchDatastore_Task, 'duration_secs': 0.019564} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.114470] env[62569]: DEBUG oslo_concurrency.lockutils [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.114850] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1013.115245] env[62569]: DEBUG oslo_concurrency.lockutils [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.115545] env[62569]: DEBUG oslo_concurrency.lockutils [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.115945] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1013.116898] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9af8b9f0-cc90-44df-9a57-0be28f765368 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.123030] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Updating instance 'e4eadcdf-a04a-4255-ba1c-fe20156c655f' progress to 83 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1013.135986] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1013.136224] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1013.137016] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d1e868a-a4ed-4a05-a47f-7930608ff697 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.143451] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 1013.143451] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5274b34d-1315-31cf-fca7-ae0d3bbf505f" [ 1013.143451] env[62569]: _type = "Task" [ 1013.143451] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.151724] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5274b34d-1315-31cf-fca7-ae0d3bbf505f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.250833] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a9509f04-8449-4089-82ca-a62fd0719f39 tempest-ServerDiskConfigTestJSON-1210343873 tempest-ServerDiskConfigTestJSON-1210343873-project-member] Lock "4df2078e-d7ce-4a3f-904b-f3303071ffff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.791s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.564543] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250602, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.631504] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1013.633033] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2b785a65-8152-46f6-b2b7-ac8c255f97fa {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.642431] env[62569]: DEBUG oslo_vmware.api [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1013.642431] env[62569]: value = "task-1250604" [ 1013.642431] env[62569]: _type = "Task" [ 1013.642431] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.665592] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5274b34d-1315-31cf-fca7-ae0d3bbf505f, 'name': SearchDatastore_Task, 'duration_secs': 0.012707} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.666109] env[62569]: DEBUG oslo_vmware.api [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250604, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.667443] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db53af17-120a-4a8d-9ab8-769621313b47 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.675669] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 1013.675669] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52296c01-2028-a49d-dfb3-a34fb21bad51" [ 1013.675669] env[62569]: _type = "Task" [ 1013.675669] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.686206] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52296c01-2028-a49d-dfb3-a34fb21bad51, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.063609] env[62569]: DEBUG oslo_vmware.api [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250602, 'name': PowerOnVM_Task, 'duration_secs': 0.697379} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.063959] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1014.064232] env[62569]: INFO nova.compute.manager [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Took 6.99 seconds to spawn the instance on the hypervisor. [ 1014.064462] env[62569]: DEBUG nova.compute.manager [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1014.065391] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54b2565c-e0f7-4f19-9957-6ac595c8b49c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.153991] env[62569]: DEBUG oslo_vmware.api [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250604, 'name': PowerOnVM_Task} progress is 84%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.189113] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52296c01-2028-a49d-dfb3-a34fb21bad51, 'name': SearchDatastore_Task, 'duration_secs': 0.011369} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.191305] env[62569]: DEBUG oslo_concurrency.lockutils [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.191753] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 0be076c2-176f-47a9-9489-2f7dfe6d1ce3/0be076c2-176f-47a9-9489-2f7dfe6d1ce3.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1014.191935] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cef5e120-460c-4448-95e2-9325ec3a8fc1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.200275] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 1014.200275] env[62569]: value = "task-1250605" [ 1014.200275] env[62569]: _type = "Task" [ 1014.200275] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.210677] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250605, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.258587] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "4bf2f0fb-420f-4c13-97c1-974af7c49658" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.258872] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "4bf2f0fb-420f-4c13-97c1-974af7c49658" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.259120] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "4bf2f0fb-420f-4c13-97c1-974af7c49658-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.259300] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "4bf2f0fb-420f-4c13-97c1-974af7c49658-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.259493] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "4bf2f0fb-420f-4c13-97c1-974af7c49658-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.261763] env[62569]: INFO nova.compute.manager [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Terminating instance [ 1014.585596] env[62569]: INFO nova.compute.manager [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Took 17.43 seconds to build instance. [ 1014.654034] env[62569]: DEBUG oslo_vmware.api [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250604, 'name': PowerOnVM_Task, 'duration_secs': 0.607431} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.654284] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1014.654490] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7afe3f27-a986-4c71-9d3e-a942359a7174 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Updating instance 'e4eadcdf-a04a-4255-ba1c-fe20156c655f' progress to 100 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1014.712840] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250605, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.765880] env[62569]: DEBUG nova.compute.manager [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1014.766142] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1014.767082] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa365ea7-26be-4819-b015-e880e5297653 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.778975] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1014.779386] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-69d74f24-0211-413b-9d1f-e3c37f60fb5c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.789366] env[62569]: DEBUG oslo_vmware.api [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1014.789366] env[62569]: value = "task-1250606" [ 1014.789366] env[62569]: _type = "Task" [ 1014.789366] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.800321] env[62569]: DEBUG oslo_vmware.api [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250606, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.087675] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cef4ef28-b84a-4932-9d92-0636065f26fe tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "c896f38c-f313-491b-86c5-a785721667d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.946s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.214149] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250605, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.718791} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.214939] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 0be076c2-176f-47a9-9489-2f7dfe6d1ce3/0be076c2-176f-47a9-9489-2f7dfe6d1ce3.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1015.215247] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1015.215500] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-020f8e4c-b012-41a2-96a4-3f65e5889007 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.225666] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 1015.225666] env[62569]: value = "task-1250607" [ 1015.225666] env[62569]: _type = "Task" [ 1015.225666] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.238335] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250607, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.300063] env[62569]: DEBUG oslo_vmware.api [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250606, 'name': PowerOffVM_Task, 'duration_secs': 0.492898} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.300063] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1015.300063] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1015.300063] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2e0e4104-19d1-433e-aa75-8fd4800c105c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.379419] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1015.379687] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1015.379905] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Deleting the datastore file [datastore2] 4bf2f0fb-420f-4c13-97c1-974af7c49658 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1015.380200] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f16ec545-55eb-4dd4-bf3e-7bb94f14ed9f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.388309] env[62569]: DEBUG oslo_vmware.api [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1015.388309] env[62569]: value = "task-1250609" [ 1015.388309] env[62569]: _type = "Task" [ 1015.388309] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.399408] env[62569]: DEBUG oslo_vmware.api [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250609, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.428517] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d398513-c562-4294-a1b3-ad3ef5085a85 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.435875] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cffdcb07-968e-47cb-a274-61b53c45e95e tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Suspending the VM {{(pid=62569) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1015.436776] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-599154e3-88c3-460a-be4a-691b1351c5e6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.443562] env[62569]: DEBUG oslo_vmware.api [None req-cffdcb07-968e-47cb-a274-61b53c45e95e tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1015.443562] env[62569]: value = "task-1250610" [ 1015.443562] env[62569]: _type = "Task" [ 1015.443562] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.453543] env[62569]: DEBUG oslo_vmware.api [None req-cffdcb07-968e-47cb-a274-61b53c45e95e tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250610, 'name': SuspendVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.739052] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250607, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.118219} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.739052] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1015.739448] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4efabcf-b4dc-45e2-9231-b7c719967988 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.761327] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 0be076c2-176f-47a9-9489-2f7dfe6d1ce3/0be076c2-176f-47a9-9489-2f7dfe6d1ce3.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1015.762492] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d8e0aa6-eaf1-4349-879e-7b09f4f49729 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.783070] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 1015.783070] env[62569]: value = "task-1250611" [ 1015.783070] env[62569]: _type = "Task" [ 1015.783070] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.792058] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250611, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.899042] env[62569]: DEBUG oslo_vmware.api [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250609, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.436985} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.899184] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1015.899421] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1015.899633] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1015.899817] env[62569]: INFO nova.compute.manager [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1015.900111] env[62569]: DEBUG oslo.service.loopingcall [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1015.900317] env[62569]: DEBUG nova.compute.manager [-] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1015.900406] env[62569]: DEBUG nova.network.neutron [-] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1015.958236] env[62569]: DEBUG oslo_vmware.api [None req-cffdcb07-968e-47cb-a274-61b53c45e95e tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250610, 'name': SuspendVM_Task} progress is 20%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.294072] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250611, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.367231] env[62569]: DEBUG nova.compute.manager [req-b963767d-c4f1-4118-9f65-056e845ba21b req-fc7ad4ea-70ee-49ca-a988-349e0ec7fc2d service nova] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Received event network-vif-deleted-3555d6f1-9762-4670-9cf9-d5d540e928a2 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1016.367231] env[62569]: INFO nova.compute.manager [req-b963767d-c4f1-4118-9f65-056e845ba21b req-fc7ad4ea-70ee-49ca-a988-349e0ec7fc2d service nova] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Neutron deleted interface 3555d6f1-9762-4670-9cf9-d5d540e928a2; detaching it from the instance and deleting it from the info cache [ 1016.367231] env[62569]: DEBUG nova.network.neutron [req-b963767d-c4f1-4118-9f65-056e845ba21b req-fc7ad4ea-70ee-49ca-a988-349e0ec7fc2d service nova] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.455283] env[62569]: DEBUG oslo_vmware.api [None req-cffdcb07-968e-47cb-a274-61b53c45e95e tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250610, 'name': SuspendVM_Task} progress is 75%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.799812] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250611, 'name': ReconfigVM_Task, 'duration_secs': 0.744035} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.804370] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 0be076c2-176f-47a9-9489-2f7dfe6d1ce3/0be076c2-176f-47a9-9489-2f7dfe6d1ce3.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1016.806641] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-283c2b47-4dd2-4752-8456-d77b87569de4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.814580] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 1016.814580] env[62569]: value = "task-1250613" [ 1016.814580] env[62569]: _type = "Task" [ 1016.814580] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.830026] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250613, 'name': Rename_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.837491] env[62569]: DEBUG nova.network.neutron [-] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.869505] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-89a82262-b0d9-4202-bbd5-42d9bd2476ce {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.886666] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12df6d94-6dc6-45c6-9ef9-2f1f0180df88 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.920264] env[62569]: DEBUG nova.compute.manager [req-b963767d-c4f1-4118-9f65-056e845ba21b req-fc7ad4ea-70ee-49ca-a988-349e0ec7fc2d service nova] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Detach interface failed, port_id=3555d6f1-9762-4670-9cf9-d5d540e928a2, reason: Instance 4bf2f0fb-420f-4c13-97c1-974af7c49658 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1016.955285] env[62569]: DEBUG oslo_vmware.api [None req-cffdcb07-968e-47cb-a274-61b53c45e95e tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250610, 'name': SuspendVM_Task, 'duration_secs': 1.324472} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.955607] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cffdcb07-968e-47cb-a274-61b53c45e95e tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Suspended the VM {{(pid=62569) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1016.955798] env[62569]: DEBUG nova.compute.manager [None req-cffdcb07-968e-47cb-a274-61b53c45e95e tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1016.956596] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c6a07b2-42dd-4c56-84e0-878108cbfb43 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.283036] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f5187d19-3a8d-4f3c-bd21-5d5d7529e456 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.284843] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f5187d19-3a8d-4f3c-bd21-5d5d7529e456 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.284843] env[62569]: DEBUG nova.compute.manager [None req-f5187d19-3a8d-4f3c-bd21-5d5d7529e456 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Going to confirm migration 4 {{(pid=62569) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5156}} [ 1017.324779] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250613, 'name': Rename_Task, 'duration_secs': 0.132498} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.324984] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1017.325258] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-34a6842c-7c87-46a1-aada-ab0c95b905b6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.331619] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 1017.331619] env[62569]: value = "task-1250614" [ 1017.331619] env[62569]: _type = "Task" [ 1017.331619] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.339277] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250614, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.339669] env[62569]: INFO nova.compute.manager [-] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Took 1.44 seconds to deallocate network for instance. [ 1017.846036] env[62569]: DEBUG oslo_vmware.api [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250614, 'name': PowerOnVM_Task, 'duration_secs': 0.463879} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.846036] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1017.846036] env[62569]: INFO nova.compute.manager [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Took 8.49 seconds to spawn the instance on the hypervisor. [ 1017.846036] env[62569]: DEBUG nova.compute.manager [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1017.846036] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35851ba9-237f-4f26-bb8b-23471b5df701 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.847688] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.847933] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.848160] env[62569]: DEBUG nova.objects.instance [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lazy-loading 'resources' on Instance uuid 4bf2f0fb-420f-4c13-97c1-974af7c49658 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1017.890651] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f5187d19-3a8d-4f3c-bd21-5d5d7529e456 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.890876] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f5187d19-3a8d-4f3c-bd21-5d5d7529e456 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.891105] env[62569]: DEBUG nova.network.neutron [None req-f5187d19-3a8d-4f3c-bd21-5d5d7529e456 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1017.891304] env[62569]: DEBUG nova.objects.instance [None req-f5187d19-3a8d-4f3c-bd21-5d5d7529e456 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lazy-loading 'info_cache' on Instance uuid e4eadcdf-a04a-4255-ba1c-fe20156c655f {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1017.962916] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "c896f38c-f313-491b-86c5-a785721667d7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.963212] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "c896f38c-f313-491b-86c5-a785721667d7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.963426] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "c896f38c-f313-491b-86c5-a785721667d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.963607] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "c896f38c-f313-491b-86c5-a785721667d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.963768] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "c896f38c-f313-491b-86c5-a785721667d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.969038] env[62569]: INFO nova.compute.manager [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Terminating instance [ 1018.362923] env[62569]: INFO nova.compute.manager [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Took 16.35 seconds to build instance. [ 1018.476409] env[62569]: DEBUG nova.compute.manager [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1018.476409] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1018.476409] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5ba353d-f69d-4075-b400-814a3af28efd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.490267] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1018.490267] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ae41d4c5-ea55-40da-9de5-0425db7e4d30 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.506771] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a341f0aa-d242-487a-8cd8-2abcf1c0454f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.514128] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9a0f762-da13-45e5-ac1c-fbaf47e7ebb5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.552451] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bcf7fed-8993-48d4-9fc4-daa4d7f2e0de {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.565271] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b2ce6d-d6f0-4694-ae35-f5749ffc1019 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.569770] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1018.569978] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1018.570174] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Deleting the datastore file [datastore2] c896f38c-f313-491b-86c5-a785721667d7 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1018.571491] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ff1a287a-0c5a-42b8-92ed-874857ddfbab {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.588221] env[62569]: DEBUG nova.compute.provider_tree [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1018.593032] env[62569]: DEBUG oslo_vmware.api [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1018.593032] env[62569]: value = "task-1250616" [ 1018.593032] env[62569]: _type = "Task" [ 1018.593032] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.599860] env[62569]: DEBUG oslo_vmware.api [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250616, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.864945] env[62569]: DEBUG oslo_concurrency.lockutils [None req-39c9fb6b-45d6-4aec-85ba-d6f7effa1a2a tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Lock "0be076c2-176f-47a9-9489-2f7dfe6d1ce3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.863s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.059218] env[62569]: INFO nova.compute.manager [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Rescuing [ 1019.059538] env[62569]: DEBUG oslo_concurrency.lockutils [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquiring lock "refresh_cache-0be076c2-176f-47a9-9489-2f7dfe6d1ce3" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.059674] env[62569]: DEBUG oslo_concurrency.lockutils [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquired lock "refresh_cache-0be076c2-176f-47a9-9489-2f7dfe6d1ce3" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.059848] env[62569]: DEBUG nova.network.neutron [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1019.107025] env[62569]: DEBUG oslo_vmware.api [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250616, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.198057} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.107025] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1019.107025] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1019.107025] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1019.107025] env[62569]: INFO nova.compute.manager [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: c896f38c-f313-491b-86c5-a785721667d7] Took 0.63 seconds to destroy the instance on the hypervisor. [ 1019.107025] env[62569]: DEBUG oslo.service.loopingcall [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1019.107025] env[62569]: DEBUG nova.compute.manager [-] [instance: c896f38c-f313-491b-86c5-a785721667d7] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1019.107025] env[62569]: DEBUG nova.network.neutron [-] [instance: c896f38c-f313-491b-86c5-a785721667d7] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1019.118328] env[62569]: ERROR nova.scheduler.client.report [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [req-0ced5cd6-911d-4128-900b-0205819ae097] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID fa06556a-5785-4014-b8bd-bc240a0cf716. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-0ced5cd6-911d-4128-900b-0205819ae097"}]} [ 1019.150731] env[62569]: DEBUG nova.scheduler.client.report [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Refreshing inventories for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1019.167353] env[62569]: DEBUG nova.scheduler.client.report [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Updating ProviderTree inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1019.167626] env[62569]: DEBUG nova.compute.provider_tree [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1019.181690] env[62569]: DEBUG nova.scheduler.client.report [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Refreshing aggregate associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, aggregates: None {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1019.209543] env[62569]: DEBUG nova.scheduler.client.report [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Refreshing trait associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1019.287450] env[62569]: DEBUG nova.network.neutron [None req-f5187d19-3a8d-4f3c-bd21-5d5d7529e456 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Updating instance_info_cache with network_info: [{"id": "39e71a4f-3b81-4b34-8cc2-2a5352eca4ce", "address": "fa:16:3e:4b:e2:a8", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap39e71a4f-3b", "ovs_interfaceid": "39e71a4f-3b81-4b34-8cc2-2a5352eca4ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.360979] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8cb3980-c60d-4134-9379-a181cc8f8e06 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.368479] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-687d7bef-736e-49e0-8041-617072d0092a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.400876] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7e1407f-7c05-4a7b-adba-10c39e6a24ac {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.408638] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21a75de4-ab3e-42a3-a8bd-6117ed48ff8b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.424657] env[62569]: DEBUG nova.compute.provider_tree [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1019.442949] env[62569]: DEBUG nova.compute.manager [req-38ba2b74-1edb-4426-8b30-48d5319464a9 req-97161f07-e07d-413f-a7a5-aa2d74575e05 service nova] [instance: c896f38c-f313-491b-86c5-a785721667d7] Received event network-vif-deleted-c2614824-b8a7-4e77-a46e-c9ac6a5dff41 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1019.443283] env[62569]: INFO nova.compute.manager [req-38ba2b74-1edb-4426-8b30-48d5319464a9 req-97161f07-e07d-413f-a7a5-aa2d74575e05 service nova] [instance: c896f38c-f313-491b-86c5-a785721667d7] Neutron deleted interface c2614824-b8a7-4e77-a46e-c9ac6a5dff41; detaching it from the instance and deleting it from the info cache [ 1019.443409] env[62569]: DEBUG nova.network.neutron [req-38ba2b74-1edb-4426-8b30-48d5319464a9 req-97161f07-e07d-413f-a7a5-aa2d74575e05 service nova] [instance: c896f38c-f313-491b-86c5-a785721667d7] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.773730] env[62569]: DEBUG nova.network.neutron [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Updating instance_info_cache with network_info: [{"id": "2e8ed1b5-2a5a-4084-b227-177a876c7cf0", "address": "fa:16:3e:93:fd:07", "network": {"id": "73711c83-88d7-490c-bc06-44e6ed11339c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1710731049-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "44464561dec24f209ec4493ac856679b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "05b1253d-2b87-4158-9ff1-dafcf829f11f", "external-id": "nsx-vlan-transportzone-55", "segmentation_id": 55, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e8ed1b5-2a", "ovs_interfaceid": "2e8ed1b5-2a5a-4084-b227-177a876c7cf0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.789857] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f5187d19-3a8d-4f3c-bd21-5d5d7529e456 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "refresh_cache-e4eadcdf-a04a-4255-ba1c-fe20156c655f" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.790157] env[62569]: DEBUG nova.objects.instance [None req-f5187d19-3a8d-4f3c-bd21-5d5d7529e456 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lazy-loading 'migration_context' on Instance uuid e4eadcdf-a04a-4255-ba1c-fe20156c655f {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1019.930965] env[62569]: DEBUG nova.network.neutron [-] [instance: c896f38c-f313-491b-86c5-a785721667d7] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.945432] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-576fbc60-fead-44b4-8c55-f6026dc8d5b9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.957072] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd3fa3bf-8801-4acf-8a94-9a8d1aa899d9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.968470] env[62569]: DEBUG nova.scheduler.client.report [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Updated inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with generation 141 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1019.968687] env[62569]: DEBUG nova.compute.provider_tree [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Updating resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 generation from 141 to 142 during operation: update_inventory {{(pid=62569) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1019.968870] env[62569]: DEBUG nova.compute.provider_tree [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1019.995665] env[62569]: DEBUG nova.compute.manager [req-38ba2b74-1edb-4426-8b30-48d5319464a9 req-97161f07-e07d-413f-a7a5-aa2d74575e05 service nova] [instance: c896f38c-f313-491b-86c5-a785721667d7] Detach interface failed, port_id=c2614824-b8a7-4e77-a46e-c9ac6a5dff41, reason: Instance c896f38c-f313-491b-86c5-a785721667d7 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1020.276897] env[62569]: DEBUG oslo_concurrency.lockutils [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Releasing lock "refresh_cache-0be076c2-176f-47a9-9489-2f7dfe6d1ce3" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.293554] env[62569]: DEBUG nova.objects.base [None req-f5187d19-3a8d-4f3c-bd21-5d5d7529e456 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62569) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1020.295886] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb0fed8-a8a6-4fc5-ad67-fa09d6e3465d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.319615] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e27314af-7f01-43fa-a3f4-191aa84ab2e9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.325769] env[62569]: DEBUG oslo_vmware.api [None req-f5187d19-3a8d-4f3c-bd21-5d5d7529e456 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1020.325769] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d43df9-4d5b-366c-3500-83b0fe23b07f" [ 1020.325769] env[62569]: _type = "Task" [ 1020.325769] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.334140] env[62569]: DEBUG oslo_vmware.api [None req-f5187d19-3a8d-4f3c-bd21-5d5d7529e456 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d43df9-4d5b-366c-3500-83b0fe23b07f, 'name': SearchDatastore_Task} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.335165] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f5187d19-3a8d-4f3c-bd21-5d5d7529e456 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.436688] env[62569]: INFO nova.compute.manager [-] [instance: c896f38c-f313-491b-86c5-a785721667d7] Took 1.33 seconds to deallocate network for instance. [ 1020.474506] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.626s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.476874] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f5187d19-3a8d-4f3c-bd21-5d5d7529e456 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.142s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.502365] env[62569]: INFO nova.scheduler.client.report [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Deleted allocations for instance 4bf2f0fb-420f-4c13-97c1-974af7c49658 [ 1020.941996] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.018129] env[62569]: DEBUG oslo_concurrency.lockutils [None req-a8acb798-125b-4a6a-8339-4df127fc65e5 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "4bf2f0fb-420f-4c13-97c1-974af7c49658" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.759s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.149990] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f40ef0be-05bc-4935-bd33-c17d7eed6f6c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.158051] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f8e9a6-d644-494a-aed7-aab09e87295a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.194517] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32c7cbd6-f8ed-401e-9db8-bd21c530740a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.202967] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f546c33e-2e04-4bc5-aea0-8d0b72e0fcc2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.217147] env[62569]: DEBUG nova.compute.provider_tree [None req-f5187d19-3a8d-4f3c-bd21-5d5d7529e456 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1021.495247] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "2ce7e7ab-8a07-468f-9f9d-643d814981bd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.495557] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "2ce7e7ab-8a07-468f-9f9d-643d814981bd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.495773] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "2ce7e7ab-8a07-468f-9f9d-643d814981bd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.495960] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "2ce7e7ab-8a07-468f-9f9d-643d814981bd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.496150] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "2ce7e7ab-8a07-468f-9f9d-643d814981bd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.498482] env[62569]: INFO nova.compute.manager [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Terminating instance [ 1021.590712] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ad2a6d77-0fcb-4696-a125-7e26589809f0 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "d56faf39-6caf-4489-98d7-342fc5fb7d40" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.590948] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ad2a6d77-0fcb-4696-a125-7e26589809f0 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "d56faf39-6caf-4489-98d7-342fc5fb7d40" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.720306] env[62569]: DEBUG nova.scheduler.client.report [None req-f5187d19-3a8d-4f3c-bd21-5d5d7529e456 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1021.830568] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1021.830852] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aa706b67-1a83-408b-bdc2-c488184017cc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.838634] env[62569]: DEBUG oslo_vmware.api [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 1021.838634] env[62569]: value = "task-1250617" [ 1021.838634] env[62569]: _type = "Task" [ 1021.838634] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.846974] env[62569]: DEBUG oslo_vmware.api [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250617, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.003232] env[62569]: DEBUG nova.compute.manager [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1022.003601] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1022.004470] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4ab42e2-cf23-44f9-b698-3c7e7969c3eb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.012186] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1022.012439] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-29154879-9477-4cb8-819b-0a152400c006 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.019062] env[62569]: DEBUG oslo_vmware.api [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1022.019062] env[62569]: value = "task-1250618" [ 1022.019062] env[62569]: _type = "Task" [ 1022.019062] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.026826] env[62569]: DEBUG oslo_vmware.api [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250618, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.094954] env[62569]: DEBUG nova.compute.utils [None req-ad2a6d77-0fcb-4696-a125-7e26589809f0 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1022.350594] env[62569]: DEBUG oslo_vmware.api [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250617, 'name': PowerOffVM_Task, 'duration_secs': 0.318259} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.350881] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1022.351701] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b875713e-0675-4cdc-896e-10a65feaa7f5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.369711] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ff72baf-1d3a-466b-8abc-c9dc6f250b96 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.398751] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1022.399021] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-96192e30-9ebb-42c0-96e9-848a20ee3b2d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.405197] env[62569]: DEBUG oslo_vmware.api [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 1022.405197] env[62569]: value = "task-1250619" [ 1022.405197] env[62569]: _type = "Task" [ 1022.405197] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.412237] env[62569]: DEBUG oslo_vmware.api [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250619, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.530424] env[62569]: DEBUG oslo_vmware.api [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250618, 'name': PowerOffVM_Task, 'duration_secs': 0.181279} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.530700] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1022.530924] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1022.531216] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b2359094-9b78-4ea5-b110-2c9eee5785a3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.589063] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1022.589223] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1022.589398] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Deleting the datastore file [datastore1] 2ce7e7ab-8a07-468f-9f9d-643d814981bd {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1022.589684] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ee0259ad-d926-44cc-8f16-192bc54c9e31 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.595575] env[62569]: DEBUG oslo_vmware.api [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1022.595575] env[62569]: value = "task-1250621" [ 1022.595575] env[62569]: _type = "Task" [ 1022.595575] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.599441] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ad2a6d77-0fcb-4696-a125-7e26589809f0 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "d56faf39-6caf-4489-98d7-342fc5fb7d40" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.606065] env[62569]: DEBUG oslo_vmware.api [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250621, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.731330] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f5187d19-3a8d-4f3c-bd21-5d5d7529e456 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.254s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.734642] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.793s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.735053] env[62569]: DEBUG nova.objects.instance [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lazy-loading 'resources' on Instance uuid c896f38c-f313-491b-86c5-a785721667d7 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1022.915941] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] VM already powered off {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1022.916180] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1022.916430] env[62569]: DEBUG oslo_concurrency.lockutils [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.916582] env[62569]: DEBUG oslo_concurrency.lockutils [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.916763] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1022.917032] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-de2d0cfc-b320-4995-b9d9-99a60fcdc284 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.924904] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1022.925100] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1022.925774] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f48e3c88-a1ea-4ffd-9ee1-c42da52fc010 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.930744] env[62569]: DEBUG oslo_vmware.api [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 1022.930744] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a4ab86-f628-c96a-73eb-2451c4985e20" [ 1022.930744] env[62569]: _type = "Task" [ 1022.930744] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.938324] env[62569]: DEBUG oslo_vmware.api [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a4ab86-f628-c96a-73eb-2451c4985e20, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.106346] env[62569]: DEBUG oslo_vmware.api [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250621, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141242} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.106638] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1023.106834] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1023.107028] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1023.107216] env[62569]: INFO nova.compute.manager [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1023.107466] env[62569]: DEBUG oslo.service.loopingcall [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1023.107669] env[62569]: DEBUG nova.compute.manager [-] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1023.107805] env[62569]: DEBUG nova.network.neutron [-] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1023.295977] env[62569]: INFO nova.scheduler.client.report [None req-f5187d19-3a8d-4f3c-bd21-5d5d7529e456 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Deleted allocation for migration a7b12128-5e18-4c5a-a47e-72831dd1bbce [ 1023.368101] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49b3da7b-b4d1-48b2-b2c3-2a93b70e0bb0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.374788] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e45dc8c-27c6-4340-8505-748d28cf7da3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.410068] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6498408-3670-4710-a6f8-81a37986929f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.413552] env[62569]: DEBUG nova.compute.manager [req-4567ed5c-e0ab-4416-b209-1e4539f6d6d7 req-88e2dfc2-1a93-474d-acc0-f6521679d717 service nova] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Received event network-vif-deleted-69a7863e-3e3c-4910-be84-9e299a18442e {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1023.413757] env[62569]: INFO nova.compute.manager [req-4567ed5c-e0ab-4416-b209-1e4539f6d6d7 req-88e2dfc2-1a93-474d-acc0-f6521679d717 service nova] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Neutron deleted interface 69a7863e-3e3c-4910-be84-9e299a18442e; detaching it from the instance and deleting it from the info cache [ 1023.413936] env[62569]: DEBUG nova.network.neutron [req-4567ed5c-e0ab-4416-b209-1e4539f6d6d7 req-88e2dfc2-1a93-474d-acc0-f6521679d717 service nova] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.421489] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51c03948-8392-48b8-8f25-799c0bcc861a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.437055] env[62569]: DEBUG nova.compute.provider_tree [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1023.447076] env[62569]: DEBUG oslo_vmware.api [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a4ab86-f628-c96a-73eb-2451c4985e20, 'name': SearchDatastore_Task, 'duration_secs': 0.008572} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.448369] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a630771a-b332-4f24-a6d0-1d21539d2672 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.453583] env[62569]: DEBUG oslo_vmware.api [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 1023.453583] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ffcd39-83b4-f0e9-0b18-a66f14c7424f" [ 1023.453583] env[62569]: _type = "Task" [ 1023.453583] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.460682] env[62569]: DEBUG oslo_vmware.api [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ffcd39-83b4-f0e9-0b18-a66f14c7424f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.664925] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ad2a6d77-0fcb-4696-a125-7e26589809f0 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "d56faf39-6caf-4489-98d7-342fc5fb7d40" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.665185] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ad2a6d77-0fcb-4696-a125-7e26589809f0 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "d56faf39-6caf-4489-98d7-342fc5fb7d40" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.665430] env[62569]: INFO nova.compute.manager [None req-ad2a6d77-0fcb-4696-a125-7e26589809f0 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Attaching volume 50b6d560-333e-4f15-a43e-84d3f4c4f0f2 to /dev/sdb [ 1023.696032] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b5bca84-4003-4651-bab8-64c67ee54737 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.703637] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bf57fdc-6ca5-4afb-a8c8-bb837b4b8946 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.717494] env[62569]: DEBUG nova.virt.block_device [None req-ad2a6d77-0fcb-4696-a125-7e26589809f0 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Updating existing volume attachment record: 9d71a6f1-f018-4e06-b7b4-f90273009fee {{(pid=62569) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1023.805739] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f5187d19-3a8d-4f3c-bd21-5d5d7529e456 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.522s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.874674] env[62569]: DEBUG nova.network.neutron [-] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.917185] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c42a2adf-ec67-4f36-b0bc-1e562f31936d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.927300] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-432cd962-d11e-41d6-8d0e-1dc1e0556ab0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.955965] env[62569]: DEBUG nova.compute.manager [req-4567ed5c-e0ab-4416-b209-1e4539f6d6d7 req-88e2dfc2-1a93-474d-acc0-f6521679d717 service nova] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Detach interface failed, port_id=69a7863e-3e3c-4910-be84-9e299a18442e, reason: Instance 2ce7e7ab-8a07-468f-9f9d-643d814981bd could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1023.957136] env[62569]: ERROR nova.scheduler.client.report [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [req-a89681fc-d353-410f-ae50-72f75ed35287] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID fa06556a-5785-4014-b8bd-bc240a0cf716. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a89681fc-d353-410f-ae50-72f75ed35287"}]} [ 1023.968730] env[62569]: DEBUG oslo_vmware.api [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ffcd39-83b4-f0e9-0b18-a66f14c7424f, 'name': SearchDatastore_Task, 'duration_secs': 0.013156} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.968730] env[62569]: DEBUG oslo_concurrency.lockutils [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.968730] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 0be076c2-176f-47a9-9489-2f7dfe6d1ce3/f77800cf-af0d-4e9c-b312-2c59488f2c7c-rescue.vmdk. {{(pid=62569) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1023.968932] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-298ae200-8f45-4c4c-bf5c-33b831e3bee0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.974784] env[62569]: DEBUG oslo_vmware.api [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 1023.974784] env[62569]: value = "task-1250623" [ 1023.974784] env[62569]: _type = "Task" [ 1023.974784] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.978842] env[62569]: DEBUG nova.scheduler.client.report [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Refreshing inventories for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1023.985442] env[62569]: DEBUG oslo_vmware.api [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250623, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.992470] env[62569]: DEBUG nova.scheduler.client.report [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Updating ProviderTree inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1023.992801] env[62569]: DEBUG nova.compute.provider_tree [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1024.004467] env[62569]: DEBUG nova.scheduler.client.report [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Refreshing aggregate associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, aggregates: None {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1024.022895] env[62569]: DEBUG nova.scheduler.client.report [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Refreshing trait associations for resource provider fa06556a-5785-4014-b8bd-bc240a0cf716, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62569) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1024.133425] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aab20e5-6b62-435d-96b2-94d459a35a8d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.141030] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27ad4837-ca4d-45b6-964f-28553c754fe2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.171898] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cc74ba9-3d9e-4d74-87ab-19bf63bd215d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.179376] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eb9ef68-0a11-4475-9136-dd4fae96abef {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.193982] env[62569]: DEBUG nova.compute.provider_tree [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1024.378065] env[62569]: INFO nova.compute.manager [-] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Took 1.27 seconds to deallocate network for instance. [ 1024.484664] env[62569]: DEBUG oslo_vmware.api [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250623, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.572474] env[62569]: DEBUG oslo_concurrency.lockutils [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.572750] env[62569]: DEBUG oslo_concurrency.lockutils [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.572967] env[62569]: DEBUG oslo_concurrency.lockutils [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.573172] env[62569]: DEBUG oslo_concurrency.lockutils [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.573349] env[62569]: DEBUG oslo_concurrency.lockutils [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.575622] env[62569]: INFO nova.compute.manager [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Terminating instance [ 1024.724680] env[62569]: DEBUG nova.scheduler.client.report [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Updated inventory for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with generation 143 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1024.725064] env[62569]: DEBUG nova.compute.provider_tree [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Updating resource provider fa06556a-5785-4014-b8bd-bc240a0cf716 generation from 143 to 144 during operation: update_inventory {{(pid=62569) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1024.725301] env[62569]: DEBUG nova.compute.provider_tree [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Updating inventory in ProviderTree for provider fa06556a-5785-4014-b8bd-bc240a0cf716 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1024.884487] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.985704] env[62569]: DEBUG oslo_vmware.api [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250623, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.574887} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.985936] env[62569]: INFO nova.virt.vmwareapi.ds_util [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 0be076c2-176f-47a9-9489-2f7dfe6d1ce3/f77800cf-af0d-4e9c-b312-2c59488f2c7c-rescue.vmdk. [ 1024.986704] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bbb52ea-b322-4719-a3c3-42879a5e66e7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.012716] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 0be076c2-176f-47a9-9489-2f7dfe6d1ce3/f77800cf-af0d-4e9c-b312-2c59488f2c7c-rescue.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1025.013022] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-74583e16-1d76-4d67-abb9-9a61b8a59e94 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.030604] env[62569]: DEBUG oslo_vmware.api [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 1025.030604] env[62569]: value = "task-1250624" [ 1025.030604] env[62569]: _type = "Task" [ 1025.030604] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.038068] env[62569]: DEBUG oslo_vmware.api [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250624, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.080117] env[62569]: DEBUG nova.compute.manager [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1025.080382] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1025.081297] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-051698b2-33ab-4a1b-b6b7-1dc70d794420 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.088620] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1025.088873] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c6463dfa-cb59-4344-b32f-10ccda420c6b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.095234] env[62569]: DEBUG oslo_vmware.api [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1025.095234] env[62569]: value = "task-1250625" [ 1025.095234] env[62569]: _type = "Task" [ 1025.095234] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.102618] env[62569]: DEBUG oslo_vmware.api [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250625, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.230052] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.495s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.232863] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.348s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.233967] env[62569]: DEBUG nova.objects.instance [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lazy-loading 'resources' on Instance uuid 2ce7e7ab-8a07-468f-9f9d-643d814981bd {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1025.254409] env[62569]: INFO nova.scheduler.client.report [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Deleted allocations for instance c896f38c-f313-491b-86c5-a785721667d7 [ 1025.540250] env[62569]: DEBUG oslo_vmware.api [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250624, 'name': ReconfigVM_Task, 'duration_secs': 0.409964} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.540561] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 0be076c2-176f-47a9-9489-2f7dfe6d1ce3/f77800cf-af0d-4e9c-b312-2c59488f2c7c-rescue.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1025.541419] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa59d1d-a175-432f-97ec-750f5fab77da {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.566986] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-25192fc6-b0f3-4f0f-a6a0-90452d569c0a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.580930] env[62569]: DEBUG oslo_vmware.api [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 1025.580930] env[62569]: value = "task-1250626" [ 1025.580930] env[62569]: _type = "Task" [ 1025.580930] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.588310] env[62569]: DEBUG oslo_vmware.api [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250626, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.605377] env[62569]: DEBUG oslo_vmware.api [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250625, 'name': PowerOffVM_Task, 'duration_secs': 0.206232} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.605645] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1025.605819] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1025.606077] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-09998617-08e3-42ab-889c-29bdc07d1cc4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.674933] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1025.675239] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1025.675434] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Deleting the datastore file [datastore1] e4eadcdf-a04a-4255-ba1c-fe20156c655f {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1025.675760] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c5bdea95-70de-46ed-af84-7ffc7224fc3a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.681576] env[62569]: DEBUG oslo_vmware.api [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1025.681576] env[62569]: value = "task-1250628" [ 1025.681576] env[62569]: _type = "Task" [ 1025.681576] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.689076] env[62569]: DEBUG oslo_vmware.api [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250628, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.764082] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f4c57f79-df83-4364-bedb-e2363a5fc0ed tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "c896f38c-f313-491b-86c5-a785721667d7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.801s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.827920] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f09efd3a-4b90-40d2-aa7b-ad5c8b88e154 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.836127] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6d9ab47-6fc5-4ff7-9a55-24241cf0990e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.865694] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f4eae53-3de5-42d5-a4e4-53adde1bcc18 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.872294] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85b818f5-2e1e-4317-a6f7-b7a624a99370 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.884917] env[62569]: DEBUG nova.compute.provider_tree [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.092933] env[62569]: DEBUG oslo_vmware.api [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250626, 'name': ReconfigVM_Task, 'duration_secs': 0.158466} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.093249] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1026.093518] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e5e20b02-106f-4a36-8399-227e71a0817b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.099674] env[62569]: DEBUG oslo_vmware.api [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 1026.099674] env[62569]: value = "task-1250630" [ 1026.099674] env[62569]: _type = "Task" [ 1026.099674] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.192351] env[62569]: DEBUG oslo_vmware.api [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250628, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15413} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.192681] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1026.192919] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1026.193163] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1026.193396] env[62569]: INFO nova.compute.manager [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1026.193699] env[62569]: DEBUG oslo.service.loopingcall [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1026.193966] env[62569]: DEBUG nova.compute.manager [-] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1026.194063] env[62569]: DEBUG nova.network.neutron [-] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1026.388492] env[62569]: DEBUG nova.scheduler.client.report [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1026.613814] env[62569]: DEBUG oslo_vmware.api [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250630, 'name': PowerOnVM_Task, 'duration_secs': 0.409471} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.613814] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1026.618351] env[62569]: DEBUG nova.compute.manager [None req-be87da21-a78f-4d6a-97ce-2760d93d46f1 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1026.618351] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed4e400b-a82e-4e3f-9901-e8a07362ff5d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.693914] env[62569]: DEBUG nova.compute.manager [req-eadeab51-d777-4f49-af8d-74b4ac07a690 req-2e44eebb-b8d2-4c69-8089-f0220fb914d1 service nova] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Received event network-vif-deleted-39e71a4f-3b81-4b34-8cc2-2a5352eca4ce {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1026.694897] env[62569]: INFO nova.compute.manager [req-eadeab51-d777-4f49-af8d-74b4ac07a690 req-2e44eebb-b8d2-4c69-8089-f0220fb914d1 service nova] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Neutron deleted interface 39e71a4f-3b81-4b34-8cc2-2a5352eca4ce; detaching it from the instance and deleting it from the info cache [ 1026.695137] env[62569]: DEBUG nova.network.neutron [req-eadeab51-d777-4f49-af8d-74b4ac07a690 req-2e44eebb-b8d2-4c69-8089-f0220fb914d1 service nova] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.893763] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.661s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.912293] env[62569]: INFO nova.scheduler.client.report [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Deleted allocations for instance 2ce7e7ab-8a07-468f-9f9d-643d814981bd [ 1027.169576] env[62569]: DEBUG nova.network.neutron [-] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.198068] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ee501b26-4f3e-4118-92f3-8c1c317a98bb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.207671] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2681bb7d-d2d6-4fd5-bf36-782288155c15 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.222067] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.222332] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.239222] env[62569]: DEBUG nova.compute.manager [req-eadeab51-d777-4f49-af8d-74b4ac07a690 req-2e44eebb-b8d2-4c69-8089-f0220fb914d1 service nova] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Detach interface failed, port_id=39e71a4f-3b81-4b34-8cc2-2a5352eca4ce, reason: Instance e4eadcdf-a04a-4255-ba1c-fe20156c655f could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1027.421326] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8208f212-fddd-44c1-9c53-5361f7564ede tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "2ce7e7ab-8a07-468f-9f9d-643d814981bd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.926s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.672617] env[62569]: INFO nova.compute.manager [-] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Took 1.48 seconds to deallocate network for instance. [ 1027.725088] env[62569]: DEBUG nova.compute.manager [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1028.178859] env[62569]: DEBUG oslo_concurrency.lockutils [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.179159] env[62569]: DEBUG oslo_concurrency.lockutils [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.179361] env[62569]: DEBUG oslo_concurrency.lockutils [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.196198] env[62569]: INFO nova.scheduler.client.report [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Deleted allocations for instance e4eadcdf-a04a-4255-ba1c-fe20156c655f [ 1028.245028] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.245135] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.246578] env[62569]: INFO nova.compute.claims [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1028.262395] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad2a6d77-0fcb-4696-a125-7e26589809f0 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Volume attach. Driver type: vmdk {{(pid=62569) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1028.262627] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad2a6d77-0fcb-4696-a125-7e26589809f0 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269502', 'volume_id': '50b6d560-333e-4f15-a43e-84d3f4c4f0f2', 'name': 'volume-50b6d560-333e-4f15-a43e-84d3f4c4f0f2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd56faf39-6caf-4489-98d7-342fc5fb7d40', 'attached_at': '', 'detached_at': '', 'volume_id': '50b6d560-333e-4f15-a43e-84d3f4c4f0f2', 'serial': '50b6d560-333e-4f15-a43e-84d3f4c4f0f2'} {{(pid=62569) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1028.263539] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87ade87a-7b08-4b37-b9ad-c71c4ed7aed7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.279910] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1a1cc8e-1aac-45bb-aba9-2b3e26cb07cb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.305156] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad2a6d77-0fcb-4696-a125-7e26589809f0 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] volume-50b6d560-333e-4f15-a43e-84d3f4c4f0f2/volume-50b6d560-333e-4f15-a43e-84d3f4c4f0f2.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1028.305650] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3730c3b3-4d1f-46f0-96d5-1ad1bb433b31 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.324092] env[62569]: DEBUG oslo_vmware.api [None req-ad2a6d77-0fcb-4696-a125-7e26589809f0 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 1028.324092] env[62569]: value = "task-1250631" [ 1028.324092] env[62569]: _type = "Task" [ 1028.324092] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.331506] env[62569]: DEBUG oslo_vmware.api [None req-ad2a6d77-0fcb-4696-a125-7e26589809f0 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250631, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.460688] env[62569]: INFO nova.compute.manager [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Unrescuing [ 1028.461056] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquiring lock "refresh_cache-0be076c2-176f-47a9-9489-2f7dfe6d1ce3" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.461130] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquired lock "refresh_cache-0be076c2-176f-47a9-9489-2f7dfe6d1ce3" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.461345] env[62569]: DEBUG nova.network.neutron [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1028.654368] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "6e3b1490-830d-4869-b805-b83ef0cf9cc7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.654686] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "6e3b1490-830d-4869-b805-b83ef0cf9cc7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.704592] env[62569]: DEBUG oslo_concurrency.lockutils [None req-244f1783-3443-40eb-8876-cdcc87d14f56 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "e4eadcdf-a04a-4255-ba1c-fe20156c655f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.132s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.834131] env[62569]: DEBUG oslo_vmware.api [None req-ad2a6d77-0fcb-4696-a125-7e26589809f0 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250631, 'name': ReconfigVM_Task, 'duration_secs': 0.326186} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.834750] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad2a6d77-0fcb-4696-a125-7e26589809f0 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Reconfigured VM instance instance-00000062 to attach disk [datastore1] volume-50b6d560-333e-4f15-a43e-84d3f4c4f0f2/volume-50b6d560-333e-4f15-a43e-84d3f4c4f0f2.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1028.840046] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac23e6eb-5cd3-470f-9679-fae165d36171 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.853982] env[62569]: DEBUG oslo_vmware.api [None req-ad2a6d77-0fcb-4696-a125-7e26589809f0 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 1028.853982] env[62569]: value = "task-1250632" [ 1028.853982] env[62569]: _type = "Task" [ 1028.853982] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.861794] env[62569]: DEBUG oslo_vmware.api [None req-ad2a6d77-0fcb-4696-a125-7e26589809f0 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250632, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.157318] env[62569]: DEBUG nova.compute.manager [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1029.161322] env[62569]: DEBUG nova.network.neutron [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Updating instance_info_cache with network_info: [{"id": "2e8ed1b5-2a5a-4084-b227-177a876c7cf0", "address": "fa:16:3e:93:fd:07", "network": {"id": "73711c83-88d7-490c-bc06-44e6ed11339c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1710731049-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "44464561dec24f209ec4493ac856679b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "05b1253d-2b87-4158-9ff1-dafcf829f11f", "external-id": "nsx-vlan-transportzone-55", "segmentation_id": 55, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e8ed1b5-2a", "ovs_interfaceid": "2e8ed1b5-2a5a-4084-b227-177a876c7cf0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.350541] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22b9372e-302a-4173-b61c-be646d1a71e3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.360197] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28b88fd4-bbce-45fb-867e-ff97cfe66f48 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.365674] env[62569]: DEBUG oslo_vmware.api [None req-ad2a6d77-0fcb-4696-a125-7e26589809f0 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250632, 'name': ReconfigVM_Task, 'duration_secs': 0.132655} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.366271] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad2a6d77-0fcb-4696-a125-7e26589809f0 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269502', 'volume_id': '50b6d560-333e-4f15-a43e-84d3f4c4f0f2', 'name': 'volume-50b6d560-333e-4f15-a43e-84d3f4c4f0f2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd56faf39-6caf-4489-98d7-342fc5fb7d40', 'attached_at': '', 'detached_at': '', 'volume_id': '50b6d560-333e-4f15-a43e-84d3f4c4f0f2', 'serial': '50b6d560-333e-4f15-a43e-84d3f4c4f0f2'} {{(pid=62569) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1029.393171] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d5fc376-3d2f-4620-9b34-681e84008bfb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.400484] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63e024eb-63fa-4d3c-9790-a46a1f20e394 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.414846] env[62569]: DEBUG nova.compute.provider_tree [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1029.575395] env[62569]: DEBUG oslo_concurrency.lockutils [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "96167629-8777-4a1c-b564-c70d1bb59b90" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.575681] env[62569]: DEBUG oslo_concurrency.lockutils [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "96167629-8777-4a1c-b564-c70d1bb59b90" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.664195] env[62569]: DEBUG oslo_concurrency.lockutils [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Releasing lock "refresh_cache-0be076c2-176f-47a9-9489-2f7dfe6d1ce3" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.664844] env[62569]: DEBUG nova.objects.instance [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Lazy-loading 'flavor' on Instance uuid 0be076c2-176f-47a9-9489-2f7dfe6d1ce3 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.680070] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.917522] env[62569]: DEBUG nova.scheduler.client.report [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1030.077890] env[62569]: DEBUG nova.compute.manager [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1030.174273] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85f58d19-9b84-48ae-ab5a-896514e31432 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.195112] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1030.195431] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a663bd7d-ef14-43a9-ae6d-3b460dff49de {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.201393] env[62569]: DEBUG oslo_vmware.api [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 1030.201393] env[62569]: value = "task-1250633" [ 1030.201393] env[62569]: _type = "Task" [ 1030.201393] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.208902] env[62569]: DEBUG oslo_vmware.api [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250633, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.401544] env[62569]: DEBUG nova.objects.instance [None req-ad2a6d77-0fcb-4696-a125-7e26589809f0 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lazy-loading 'flavor' on Instance uuid d56faf39-6caf-4489-98d7-342fc5fb7d40 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1030.422358] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.177s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.423022] env[62569]: DEBUG nova.compute.manager [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1030.425636] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.746s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.427607] env[62569]: INFO nova.compute.claims [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1030.600039] env[62569]: DEBUG oslo_concurrency.lockutils [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.711382] env[62569]: DEBUG oslo_vmware.api [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250633, 'name': PowerOffVM_Task, 'duration_secs': 0.23687} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.711651] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1030.716925] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Reconfiguring VM instance instance-00000069 to detach disk 2001 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1030.717204] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e200806-0f5a-4779-838e-0f9722cd2223 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.734580] env[62569]: DEBUG oslo_vmware.api [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 1030.734580] env[62569]: value = "task-1250634" [ 1030.734580] env[62569]: _type = "Task" [ 1030.734580] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.743887] env[62569]: DEBUG oslo_vmware.api [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250634, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.906602] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ad2a6d77-0fcb-4696-a125-7e26589809f0 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "d56faf39-6caf-4489-98d7-342fc5fb7d40" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.241s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.934188] env[62569]: DEBUG nova.compute.utils [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1030.935704] env[62569]: DEBUG nova.compute.manager [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1030.935914] env[62569]: DEBUG nova.network.neutron [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1030.981681] env[62569]: DEBUG nova.policy [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5da4f6b7a6784a73bd3fed04b275041b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bdba6022e3d4697a336ca28ca4eccec', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 1031.137661] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e8ba9ac9-d56b-41e8-8bc6-f535081f38de tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "d56faf39-6caf-4489-98d7-342fc5fb7d40" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.137966] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e8ba9ac9-d56b-41e8-8bc6-f535081f38de tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "d56faf39-6caf-4489-98d7-342fc5fb7d40" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.246151] env[62569]: DEBUG oslo_vmware.api [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250634, 'name': ReconfigVM_Task, 'duration_secs': 0.225637} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.246151] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Reconfigured VM instance instance-00000069 to detach disk 2001 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1031.246151] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1031.246151] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-362dc55d-37b8-42be-af3c-3c3936a45541 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.251711] env[62569]: DEBUG oslo_vmware.api [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 1031.251711] env[62569]: value = "task-1250635" [ 1031.251711] env[62569]: _type = "Task" [ 1031.251711] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.256962] env[62569]: DEBUG nova.network.neutron [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Successfully created port: 04dded5b-c41d-4dd5-b9bb-354115b85a5d {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1031.262032] env[62569]: DEBUG oslo_vmware.api [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250635, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.440537] env[62569]: DEBUG nova.compute.manager [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1031.538166] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b263cfac-7ea9-43f7-b61a-f29b0c0c051d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.545891] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e29331a-9e25-4de9-b04d-bc5620184e03 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.576657] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-945e680b-3d66-4dc8-a065-99a76817f667 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.583246] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51e46c58-de6a-4d7d-8e4e-4d794f4a97ff {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.595618] env[62569]: DEBUG nova.compute.provider_tree [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1031.644574] env[62569]: INFO nova.compute.manager [None req-e8ba9ac9-d56b-41e8-8bc6-f535081f38de tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Detaching volume 50b6d560-333e-4f15-a43e-84d3f4c4f0f2 [ 1031.678715] env[62569]: INFO nova.virt.block_device [None req-e8ba9ac9-d56b-41e8-8bc6-f535081f38de tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Attempting to driver detach volume 50b6d560-333e-4f15-a43e-84d3f4c4f0f2 from mountpoint /dev/sdb [ 1031.678958] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8ba9ac9-d56b-41e8-8bc6-f535081f38de tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Volume detach. Driver type: vmdk {{(pid=62569) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1031.679168] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8ba9ac9-d56b-41e8-8bc6-f535081f38de tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269502', 'volume_id': '50b6d560-333e-4f15-a43e-84d3f4c4f0f2', 'name': 'volume-50b6d560-333e-4f15-a43e-84d3f4c4f0f2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd56faf39-6caf-4489-98d7-342fc5fb7d40', 'attached_at': '', 'detached_at': '', 'volume_id': '50b6d560-333e-4f15-a43e-84d3f4c4f0f2', 'serial': '50b6d560-333e-4f15-a43e-84d3f4c4f0f2'} {{(pid=62569) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1031.680095] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-629db225-08be-4c5c-80b5-6a66e343b3f2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.702556] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a283623f-f499-4a8b-a313-18847ac5b18e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.709390] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-839505b6-7c3e-474b-8a0f-6445e3ae753b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.728320] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d17feba9-6049-4e4d-af56-304075b5b6a0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.742283] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8ba9ac9-d56b-41e8-8bc6-f535081f38de tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] The volume has not been displaced from its original location: [datastore1] volume-50b6d560-333e-4f15-a43e-84d3f4c4f0f2/volume-50b6d560-333e-4f15-a43e-84d3f4c4f0f2.vmdk. No consolidation needed. {{(pid=62569) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1031.747366] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8ba9ac9-d56b-41e8-8bc6-f535081f38de tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Reconfiguring VM instance instance-00000062 to detach disk 2001 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1031.747613] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-941db13c-bbb8-404e-a0ee-05185a08adbc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.768833] env[62569]: DEBUG oslo_vmware.api [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250635, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.769898] env[62569]: DEBUG oslo_vmware.api [None req-e8ba9ac9-d56b-41e8-8bc6-f535081f38de tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 1031.769898] env[62569]: value = "task-1250636" [ 1031.769898] env[62569]: _type = "Task" [ 1031.769898] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.098575] env[62569]: DEBUG nova.scheduler.client.report [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1032.268873] env[62569]: DEBUG oslo_vmware.api [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250635, 'name': PowerOnVM_Task, 'duration_secs': 0.988492} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.269156] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1032.269400] env[62569]: DEBUG nova.compute.manager [None req-12547c5e-7711-409f-8d4d-638dc9712fa2 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1032.270153] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02240bc8-cae2-4116-9467-e6467d97380d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.282858] env[62569]: DEBUG oslo_vmware.api [None req-e8ba9ac9-d56b-41e8-8bc6-f535081f38de tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250636, 'name': ReconfigVM_Task, 'duration_secs': 0.215632} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.283324] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8ba9ac9-d56b-41e8-8bc6-f535081f38de tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Reconfigured VM instance instance-00000062 to detach disk 2001 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1032.288198] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01929a0b-166c-455c-b1dd-d5735a000a08 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.302504] env[62569]: DEBUG oslo_vmware.api [None req-e8ba9ac9-d56b-41e8-8bc6-f535081f38de tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 1032.302504] env[62569]: value = "task-1250637" [ 1032.302504] env[62569]: _type = "Task" [ 1032.302504] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.310932] env[62569]: DEBUG oslo_vmware.api [None req-e8ba9ac9-d56b-41e8-8bc6-f535081f38de tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250637, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.449368] env[62569]: DEBUG nova.compute.manager [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1032.474114] env[62569]: DEBUG nova.virt.hardware [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1032.474368] env[62569]: DEBUG nova.virt.hardware [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1032.474530] env[62569]: DEBUG nova.virt.hardware [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1032.474714] env[62569]: DEBUG nova.virt.hardware [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1032.474863] env[62569]: DEBUG nova.virt.hardware [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1032.475025] env[62569]: DEBUG nova.virt.hardware [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1032.475244] env[62569]: DEBUG nova.virt.hardware [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1032.475405] env[62569]: DEBUG nova.virt.hardware [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1032.475575] env[62569]: DEBUG nova.virt.hardware [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1032.475738] env[62569]: DEBUG nova.virt.hardware [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1032.475911] env[62569]: DEBUG nova.virt.hardware [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1032.476784] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bef7124-e754-499c-b29a-ff5e7564bd20 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.485257] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4eb2f15-0eff-4957-90bf-180cf487f0c3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.604533] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.179s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.605218] env[62569]: DEBUG nova.compute.manager [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1032.608243] env[62569]: DEBUG oslo_concurrency.lockutils [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.008s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.609995] env[62569]: INFO nova.compute.claims [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1032.655218] env[62569]: DEBUG nova.compute.manager [req-528249a8-efb4-42f6-b03b-a42e18fb1e7f req-4a06b0b4-8e4a-4082-b7d7-a4ae06123aa7 service nova] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Received event network-vif-plugged-04dded5b-c41d-4dd5-b9bb-354115b85a5d {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1032.655486] env[62569]: DEBUG oslo_concurrency.lockutils [req-528249a8-efb4-42f6-b03b-a42e18fb1e7f req-4a06b0b4-8e4a-4082-b7d7-a4ae06123aa7 service nova] Acquiring lock "f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.655735] env[62569]: DEBUG oslo_concurrency.lockutils [req-528249a8-efb4-42f6-b03b-a42e18fb1e7f req-4a06b0b4-8e4a-4082-b7d7-a4ae06123aa7 service nova] Lock "f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.655969] env[62569]: DEBUG oslo_concurrency.lockutils [req-528249a8-efb4-42f6-b03b-a42e18fb1e7f req-4a06b0b4-8e4a-4082-b7d7-a4ae06123aa7 service nova] Lock "f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.656200] env[62569]: DEBUG nova.compute.manager [req-528249a8-efb4-42f6-b03b-a42e18fb1e7f req-4a06b0b4-8e4a-4082-b7d7-a4ae06123aa7 service nova] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] No waiting events found dispatching network-vif-plugged-04dded5b-c41d-4dd5-b9bb-354115b85a5d {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1032.656412] env[62569]: WARNING nova.compute.manager [req-528249a8-efb4-42f6-b03b-a42e18fb1e7f req-4a06b0b4-8e4a-4082-b7d7-a4ae06123aa7 service nova] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Received unexpected event network-vif-plugged-04dded5b-c41d-4dd5-b9bb-354115b85a5d for instance with vm_state building and task_state spawning. [ 1032.717950] env[62569]: DEBUG nova.network.neutron [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Successfully updated port: 04dded5b-c41d-4dd5-b9bb-354115b85a5d {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1032.814024] env[62569]: DEBUG oslo_vmware.api [None req-e8ba9ac9-d56b-41e8-8bc6-f535081f38de tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250637, 'name': ReconfigVM_Task, 'duration_secs': 0.12953} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.814347] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8ba9ac9-d56b-41e8-8bc6-f535081f38de tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269502', 'volume_id': '50b6d560-333e-4f15-a43e-84d3f4c4f0f2', 'name': 'volume-50b6d560-333e-4f15-a43e-84d3f4c4f0f2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd56faf39-6caf-4489-98d7-342fc5fb7d40', 'attached_at': '', 'detached_at': '', 'volume_id': '50b6d560-333e-4f15-a43e-84d3f4c4f0f2', 'serial': '50b6d560-333e-4f15-a43e-84d3f4c4f0f2'} {{(pid=62569) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1033.115278] env[62569]: DEBUG nova.compute.utils [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1033.118411] env[62569]: DEBUG nova.compute.manager [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1033.118958] env[62569]: DEBUG nova.network.neutron [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1033.141723] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquiring lock "0be076c2-176f-47a9-9489-2f7dfe6d1ce3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.141987] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Lock "0be076c2-176f-47a9-9489-2f7dfe6d1ce3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.142240] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquiring lock "0be076c2-176f-47a9-9489-2f7dfe6d1ce3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.142441] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Lock "0be076c2-176f-47a9-9489-2f7dfe6d1ce3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.142671] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Lock "0be076c2-176f-47a9-9489-2f7dfe6d1ce3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.144530] env[62569]: INFO nova.compute.manager [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Terminating instance [ 1033.164520] env[62569]: DEBUG nova.policy [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6dd052494f244fd78050c5d26f497720', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2c6a4814c234d53bbd4844cd482fbae', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 1033.224057] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "refresh_cache-f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.224234] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired lock "refresh_cache-f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.224381] env[62569]: DEBUG nova.network.neutron [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1033.359574] env[62569]: DEBUG nova.objects.instance [None req-e8ba9ac9-d56b-41e8-8bc6-f535081f38de tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lazy-loading 'flavor' on Instance uuid d56faf39-6caf-4489-98d7-342fc5fb7d40 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.398190] env[62569]: DEBUG nova.network.neutron [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Successfully created port: 42cd0220-15ec-4c55-b1f3-7c583246ff5a {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1033.621769] env[62569]: DEBUG nova.compute.manager [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1033.648312] env[62569]: DEBUG nova.compute.manager [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1033.648561] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1033.649477] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1d297e0-c6f8-4e7e-81b9-5a71739d1850 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.657176] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1033.660042] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8013eac9-1549-415e-8d88-62f6448939d5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.666543] env[62569]: DEBUG oslo_vmware.api [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 1033.666543] env[62569]: value = "task-1250638" [ 1033.666543] env[62569]: _type = "Task" [ 1033.666543] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.675301] env[62569]: DEBUG oslo_vmware.api [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250638, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.738516] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b852691-c357-41da-a6f2-b5202de5795a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.746569] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1de17caf-6466-4962-84c1-07e2ae5f2342 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.780436] env[62569]: DEBUG nova.network.neutron [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1033.782793] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0e2f7ef-1d8e-4e16-aaa0-c300c3af7113 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.790939] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9251a66-451b-42d4-9612-f5ac19cf0838 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.805224] env[62569]: DEBUG nova.compute.provider_tree [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1033.923034] env[62569]: DEBUG nova.network.neutron [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Updating instance_info_cache with network_info: [{"id": "04dded5b-c41d-4dd5-b9bb-354115b85a5d", "address": "fa:16:3e:50:e2:7f", "network": {"id": "30d3bf4a-8f55-492f-a049-181e29731818", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1244949475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bdba6022e3d4697a336ca28ca4eccec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04dded5b-c4", "ovs_interfaceid": "04dded5b-c41d-4dd5-b9bb-354115b85a5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.178045] env[62569]: DEBUG oslo_vmware.api [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250638, 'name': PowerOffVM_Task, 'duration_secs': 0.188556} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.178045] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1034.178045] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1034.178259] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a7058843-c037-4044-97b0-81c62e53f456 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.241021] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1034.241452] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1034.241759] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Deleting the datastore file [datastore2] 0be076c2-176f-47a9-9489-2f7dfe6d1ce3 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1034.242219] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cc86aea6-71ee-44d1-9f0d-0c313fbabd60 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.249440] env[62569]: DEBUG oslo_vmware.api [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 1034.249440] env[62569]: value = "task-1250640" [ 1034.249440] env[62569]: _type = "Task" [ 1034.249440] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.257824] env[62569]: DEBUG oslo_vmware.api [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250640, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.309460] env[62569]: DEBUG nova.scheduler.client.report [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1034.366927] env[62569]: DEBUG oslo_concurrency.lockutils [None req-e8ba9ac9-d56b-41e8-8bc6-f535081f38de tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "d56faf39-6caf-4489-98d7-342fc5fb7d40" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.228s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.427018] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Releasing lock "refresh_cache-f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.427018] env[62569]: DEBUG nova.compute.manager [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Instance network_info: |[{"id": "04dded5b-c41d-4dd5-b9bb-354115b85a5d", "address": "fa:16:3e:50:e2:7f", "network": {"id": "30d3bf4a-8f55-492f-a049-181e29731818", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1244949475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bdba6022e3d4697a336ca28ca4eccec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04dded5b-c4", "ovs_interfaceid": "04dded5b-c41d-4dd5-b9bb-354115b85a5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1034.427018] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:50:e2:7f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '983826cf-6390-4ec6-bf97-30a1060947fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '04dded5b-c41d-4dd5-b9bb-354115b85a5d', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1034.433466] env[62569]: DEBUG oslo.service.loopingcall [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1034.433856] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1034.434231] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e5c6588c-2f80-4da7-a7ae-f9e74e9574e5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.453993] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1034.453993] env[62569]: value = "task-1250641" [ 1034.453993] env[62569]: _type = "Task" [ 1034.453993] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.461556] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250641, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.634765] env[62569]: DEBUG nova.compute.manager [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1034.663021] env[62569]: DEBUG nova.virt.hardware [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1034.663476] env[62569]: DEBUG nova.virt.hardware [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1034.663756] env[62569]: DEBUG nova.virt.hardware [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1034.664104] env[62569]: DEBUG nova.virt.hardware [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1034.664374] env[62569]: DEBUG nova.virt.hardware [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1034.664654] env[62569]: DEBUG nova.virt.hardware [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1034.665015] env[62569]: DEBUG nova.virt.hardware [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1034.665308] env[62569]: DEBUG nova.virt.hardware [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1034.665618] env[62569]: DEBUG nova.virt.hardware [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1034.665926] env[62569]: DEBUG nova.virt.hardware [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1034.666236] env[62569]: DEBUG nova.virt.hardware [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1034.667535] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd807ba2-e91a-425c-a86a-690257c63602 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.678710] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8c3f1ed-d126-406c-920e-5da20cbf84b7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.695678] env[62569]: DEBUG nova.compute.manager [req-aa00eb9e-6be9-4d31-9fa8-d40457c4973f req-5ff7e465-ccf4-4a65-b2c8-b46ac83e3f00 service nova] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Received event network-changed-04dded5b-c41d-4dd5-b9bb-354115b85a5d {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1034.695987] env[62569]: DEBUG nova.compute.manager [req-aa00eb9e-6be9-4d31-9fa8-d40457c4973f req-5ff7e465-ccf4-4a65-b2c8-b46ac83e3f00 service nova] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Refreshing instance network info cache due to event network-changed-04dded5b-c41d-4dd5-b9bb-354115b85a5d. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1034.696098] env[62569]: DEBUG oslo_concurrency.lockutils [req-aa00eb9e-6be9-4d31-9fa8-d40457c4973f req-5ff7e465-ccf4-4a65-b2c8-b46ac83e3f00 service nova] Acquiring lock "refresh_cache-f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.696248] env[62569]: DEBUG oslo_concurrency.lockutils [req-aa00eb9e-6be9-4d31-9fa8-d40457c4973f req-5ff7e465-ccf4-4a65-b2c8-b46ac83e3f00 service nova] Acquired lock "refresh_cache-f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.696412] env[62569]: DEBUG nova.network.neutron [req-aa00eb9e-6be9-4d31-9fa8-d40457c4973f req-5ff7e465-ccf4-4a65-b2c8-b46ac83e3f00 service nova] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Refreshing network info cache for port 04dded5b-c41d-4dd5-b9bb-354115b85a5d {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1034.758517] env[62569]: DEBUG oslo_vmware.api [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250640, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130702} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.758801] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1034.758994] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1034.759199] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1034.759453] env[62569]: INFO nova.compute.manager [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1034.759690] env[62569]: DEBUG oslo.service.loopingcall [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1034.759832] env[62569]: DEBUG nova.compute.manager [-] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1034.759939] env[62569]: DEBUG nova.network.neutron [-] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1034.815361] env[62569]: DEBUG oslo_concurrency.lockutils [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.206s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.815472] env[62569]: DEBUG nova.compute.manager [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1034.964853] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250641, 'name': CreateVM_Task, 'duration_secs': 0.279752} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.965117] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1034.966675] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.966675] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.966675] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1034.967020] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54345000-ca6b-4db3-a612-136a5270f3ad {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.971827] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1034.971827] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ff61a3-d116-f7bf-8fd3-e4892d2c34bc" [ 1034.971827] env[62569]: _type = "Task" [ 1034.971827] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.979615] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ff61a3-d116-f7bf-8fd3-e4892d2c34bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.980501] env[62569]: DEBUG nova.network.neutron [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Successfully updated port: 42cd0220-15ec-4c55-b1f3-7c583246ff5a {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1035.320435] env[62569]: DEBUG nova.compute.utils [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1035.321849] env[62569]: DEBUG nova.compute.manager [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1035.322101] env[62569]: DEBUG nova.network.neutron [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1035.383476] env[62569]: DEBUG oslo_concurrency.lockutils [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "d56faf39-6caf-4489-98d7-342fc5fb7d40" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.383989] env[62569]: DEBUG oslo_concurrency.lockutils [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "d56faf39-6caf-4489-98d7-342fc5fb7d40" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.384442] env[62569]: DEBUG oslo_concurrency.lockutils [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "d56faf39-6caf-4489-98d7-342fc5fb7d40-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.384809] env[62569]: DEBUG oslo_concurrency.lockutils [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "d56faf39-6caf-4489-98d7-342fc5fb7d40-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.385207] env[62569]: DEBUG oslo_concurrency.lockutils [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "d56faf39-6caf-4489-98d7-342fc5fb7d40-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.387891] env[62569]: DEBUG nova.policy [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ccb9b3f4e38e4e81819d22e781661015', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7f988acdb9cd4ca28ca8916c2ede2db5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 1035.389595] env[62569]: INFO nova.compute.manager [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Terminating instance [ 1035.450453] env[62569]: DEBUG nova.network.neutron [req-aa00eb9e-6be9-4d31-9fa8-d40457c4973f req-5ff7e465-ccf4-4a65-b2c8-b46ac83e3f00 service nova] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Updated VIF entry in instance network info cache for port 04dded5b-c41d-4dd5-b9bb-354115b85a5d. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1035.450921] env[62569]: DEBUG nova.network.neutron [req-aa00eb9e-6be9-4d31-9fa8-d40457c4973f req-5ff7e465-ccf4-4a65-b2c8-b46ac83e3f00 service nova] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Updating instance_info_cache with network_info: [{"id": "04dded5b-c41d-4dd5-b9bb-354115b85a5d", "address": "fa:16:3e:50:e2:7f", "network": {"id": "30d3bf4a-8f55-492f-a049-181e29731818", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1244949475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bdba6022e3d4697a336ca28ca4eccec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04dded5b-c4", "ovs_interfaceid": "04dded5b-c41d-4dd5-b9bb-354115b85a5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.481567] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ff61a3-d116-f7bf-8fd3-e4892d2c34bc, 'name': SearchDatastore_Task, 'duration_secs': 0.008937} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.481898] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.482149] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1035.482399] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.482607] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.482719] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1035.483239] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "refresh_cache-6e3b1490-830d-4869-b805-b83ef0cf9cc7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.483365] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired lock "refresh_cache-6e3b1490-830d-4869-b805-b83ef0cf9cc7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.483515] env[62569]: DEBUG nova.network.neutron [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1035.484461] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e6f0255d-bb05-4c1c-9a0d-f9506bdf7804 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.493414] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1035.493606] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1035.494327] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5b37b1b-8209-4dc5-a420-f8406bb3f172 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.499561] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1035.499561] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52659982-a056-361d-40c4-bb5cfe0ef140" [ 1035.499561] env[62569]: _type = "Task" [ 1035.499561] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.507176] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52659982-a056-361d-40c4-bb5cfe0ef140, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.582020] env[62569]: DEBUG nova.network.neutron [-] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.664828] env[62569]: DEBUG nova.network.neutron [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Successfully created port: d5f53bf1-303b-435e-941a-47c7ab293484 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1035.825394] env[62569]: DEBUG nova.compute.manager [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1035.893081] env[62569]: DEBUG nova.compute.manager [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1035.893332] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1035.894293] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce4a2d4f-f4ee-422e-9735-a20342d9641e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.902197] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1035.902451] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7523640e-b2e6-4da9-a14f-27c67075c732 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.908606] env[62569]: DEBUG oslo_vmware.api [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 1035.908606] env[62569]: value = "task-1250642" [ 1035.908606] env[62569]: _type = "Task" [ 1035.908606] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.916462] env[62569]: DEBUG oslo_vmware.api [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250642, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.953731] env[62569]: DEBUG oslo_concurrency.lockutils [req-aa00eb9e-6be9-4d31-9fa8-d40457c4973f req-5ff7e465-ccf4-4a65-b2c8-b46ac83e3f00 service nova] Releasing lock "refresh_cache-f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.010976] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52659982-a056-361d-40c4-bb5cfe0ef140, 'name': SearchDatastore_Task, 'duration_secs': 0.008434} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.011828] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-864f6b1a-63d1-4cc2-9c87-f41c1ceea285 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.015871] env[62569]: DEBUG nova.network.neutron [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1036.019199] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1036.019199] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ff0e56-90fe-8e63-9831-3795544bf83a" [ 1036.019199] env[62569]: _type = "Task" [ 1036.019199] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.026705] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ff0e56-90fe-8e63-9831-3795544bf83a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.081284] env[62569]: INFO nova.compute.manager [-] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Took 1.32 seconds to deallocate network for instance. [ 1036.152269] env[62569]: DEBUG nova.network.neutron [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Updating instance_info_cache with network_info: [{"id": "42cd0220-15ec-4c55-b1f3-7c583246ff5a", "address": "fa:16:3e:ef:b0:a4", "network": {"id": "884d3248-07f1-4cf7-8ceb-b94b34622384", "bridge": "br-int", "label": "tempest-ServersTestJSON-1882023778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2c6a4814c234d53bbd4844cd482fbae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42cd0220-15", "ovs_interfaceid": "42cd0220-15ec-4c55-b1f3-7c583246ff5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.418283] env[62569]: DEBUG oslo_vmware.api [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250642, 'name': PowerOffVM_Task, 'duration_secs': 0.21045} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.418564] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1036.418736] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1036.419007] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2f9aebd9-8293-4340-93db-a77788622ba9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.480018] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1036.480279] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1036.480462] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Deleting the datastore file [datastore2] d56faf39-6caf-4489-98d7-342fc5fb7d40 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1036.480740] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-190edb4e-0e42-49b0-aba3-172c715496be {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.486806] env[62569]: DEBUG oslo_vmware.api [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 1036.486806] env[62569]: value = "task-1250644" [ 1036.486806] env[62569]: _type = "Task" [ 1036.486806] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.494760] env[62569]: DEBUG oslo_vmware.api [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250644, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.528366] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52ff0e56-90fe-8e63-9831-3795544bf83a, 'name': SearchDatastore_Task, 'duration_secs': 0.00971} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.528613] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.528872] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232/f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1036.529132] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a3587516-c23e-433f-932a-2aa800053f76 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.535144] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1036.535144] env[62569]: value = "task-1250645" [ 1036.535144] env[62569]: _type = "Task" [ 1036.535144] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.542597] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250645, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.588197] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.588535] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.588785] env[62569]: DEBUG nova.objects.instance [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Lazy-loading 'resources' on Instance uuid 0be076c2-176f-47a9-9489-2f7dfe6d1ce3 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1036.654880] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Releasing lock "refresh_cache-6e3b1490-830d-4869-b805-b83ef0cf9cc7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.655247] env[62569]: DEBUG nova.compute.manager [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Instance network_info: |[{"id": "42cd0220-15ec-4c55-b1f3-7c583246ff5a", "address": "fa:16:3e:ef:b0:a4", "network": {"id": "884d3248-07f1-4cf7-8ceb-b94b34622384", "bridge": "br-int", "label": "tempest-ServersTestJSON-1882023778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2c6a4814c234d53bbd4844cd482fbae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42cd0220-15", "ovs_interfaceid": "42cd0220-15ec-4c55-b1f3-7c583246ff5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1036.655694] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ef:b0:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e9fa4744-8702-4973-b911-ee18192a3e4b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '42cd0220-15ec-4c55-b1f3-7c583246ff5a', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1036.663328] env[62569]: DEBUG oslo.service.loopingcall [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1036.663858] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1036.664119] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-27ea5db7-f591-4755-820e-8bb81e01454c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.684522] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1036.684522] env[62569]: value = "task-1250646" [ 1036.684522] env[62569]: _type = "Task" [ 1036.684522] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.692890] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250646, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.720820] env[62569]: DEBUG nova.compute.manager [req-56abd7db-adb4-4bbd-a224-619d17245d67 req-1676c48f-4033-4e9a-8c87-7d1581cf3940 service nova] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Received event network-vif-plugged-42cd0220-15ec-4c55-b1f3-7c583246ff5a {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1036.721068] env[62569]: DEBUG oslo_concurrency.lockutils [req-56abd7db-adb4-4bbd-a224-619d17245d67 req-1676c48f-4033-4e9a-8c87-7d1581cf3940 service nova] Acquiring lock "6e3b1490-830d-4869-b805-b83ef0cf9cc7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.721288] env[62569]: DEBUG oslo_concurrency.lockutils [req-56abd7db-adb4-4bbd-a224-619d17245d67 req-1676c48f-4033-4e9a-8c87-7d1581cf3940 service nova] Lock "6e3b1490-830d-4869-b805-b83ef0cf9cc7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.721466] env[62569]: DEBUG oslo_concurrency.lockutils [req-56abd7db-adb4-4bbd-a224-619d17245d67 req-1676c48f-4033-4e9a-8c87-7d1581cf3940 service nova] Lock "6e3b1490-830d-4869-b805-b83ef0cf9cc7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.721642] env[62569]: DEBUG nova.compute.manager [req-56abd7db-adb4-4bbd-a224-619d17245d67 req-1676c48f-4033-4e9a-8c87-7d1581cf3940 service nova] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] No waiting events found dispatching network-vif-plugged-42cd0220-15ec-4c55-b1f3-7c583246ff5a {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1036.721813] env[62569]: WARNING nova.compute.manager [req-56abd7db-adb4-4bbd-a224-619d17245d67 req-1676c48f-4033-4e9a-8c87-7d1581cf3940 service nova] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Received unexpected event network-vif-plugged-42cd0220-15ec-4c55-b1f3-7c583246ff5a for instance with vm_state building and task_state spawning. [ 1036.722071] env[62569]: DEBUG nova.compute.manager [req-56abd7db-adb4-4bbd-a224-619d17245d67 req-1676c48f-4033-4e9a-8c87-7d1581cf3940 service nova] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Received event network-changed-42cd0220-15ec-4c55-b1f3-7c583246ff5a {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1036.722276] env[62569]: DEBUG nova.compute.manager [req-56abd7db-adb4-4bbd-a224-619d17245d67 req-1676c48f-4033-4e9a-8c87-7d1581cf3940 service nova] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Refreshing instance network info cache due to event network-changed-42cd0220-15ec-4c55-b1f3-7c583246ff5a. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1036.722469] env[62569]: DEBUG oslo_concurrency.lockutils [req-56abd7db-adb4-4bbd-a224-619d17245d67 req-1676c48f-4033-4e9a-8c87-7d1581cf3940 service nova] Acquiring lock "refresh_cache-6e3b1490-830d-4869-b805-b83ef0cf9cc7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.722611] env[62569]: DEBUG oslo_concurrency.lockutils [req-56abd7db-adb4-4bbd-a224-619d17245d67 req-1676c48f-4033-4e9a-8c87-7d1581cf3940 service nova] Acquired lock "refresh_cache-6e3b1490-830d-4869-b805-b83ef0cf9cc7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.722770] env[62569]: DEBUG nova.network.neutron [req-56abd7db-adb4-4bbd-a224-619d17245d67 req-1676c48f-4033-4e9a-8c87-7d1581cf3940 service nova] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Refreshing network info cache for port 42cd0220-15ec-4c55-b1f3-7c583246ff5a {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1036.835542] env[62569]: DEBUG nova.compute.manager [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1036.863716] env[62569]: DEBUG nova.virt.hardware [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1036.864082] env[62569]: DEBUG nova.virt.hardware [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1036.864303] env[62569]: DEBUG nova.virt.hardware [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1036.864546] env[62569]: DEBUG nova.virt.hardware [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1036.864747] env[62569]: DEBUG nova.virt.hardware [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1036.864923] env[62569]: DEBUG nova.virt.hardware [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1036.865239] env[62569]: DEBUG nova.virt.hardware [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1036.865441] env[62569]: DEBUG nova.virt.hardware [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1036.865662] env[62569]: DEBUG nova.virt.hardware [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1036.865856] env[62569]: DEBUG nova.virt.hardware [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1036.866110] env[62569]: DEBUG nova.virt.hardware [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1036.867524] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6ee61f9-d67a-42eb-9fd1-8edeae42ab27 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.876437] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ef0d2d-1ab1-4f66-ae18-e27639de9262 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.997307] env[62569]: DEBUG oslo_vmware.api [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250644, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136014} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.997571] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1036.997868] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1036.997979] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1036.998276] env[62569]: INFO nova.compute.manager [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1036.999397] env[62569]: DEBUG oslo.service.loopingcall [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1036.999397] env[62569]: DEBUG nova.compute.manager [-] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1036.999397] env[62569]: DEBUG nova.network.neutron [-] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1037.045105] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250645, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.455065} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.045400] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232/f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1037.045627] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1037.045875] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8fa32e33-e9cf-47a0-bbb7-0e7a1aeb8bd8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.052109] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1037.052109] env[62569]: value = "task-1250647" [ 1037.052109] env[62569]: _type = "Task" [ 1037.052109] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.059478] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250647, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.194302] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250646, 'name': CreateVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.229086] env[62569]: DEBUG nova.network.neutron [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Successfully updated port: d5f53bf1-303b-435e-941a-47c7ab293484 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1037.232773] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8f9ceb8-bf98-4fd6-9dc8-72ddb95ab830 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.241516] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b2deb78-e3c1-439b-91d1-9866575bc003 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.275132] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c62ba8d2-689b-4780-887d-62f0d1d68400 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.283050] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26bb0307-d87e-4dbf-af57-44f9936d4089 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.297805] env[62569]: DEBUG nova.compute.provider_tree [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1037.464818] env[62569]: DEBUG nova.network.neutron [req-56abd7db-adb4-4bbd-a224-619d17245d67 req-1676c48f-4033-4e9a-8c87-7d1581cf3940 service nova] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Updated VIF entry in instance network info cache for port 42cd0220-15ec-4c55-b1f3-7c583246ff5a. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1037.465927] env[62569]: DEBUG nova.network.neutron [req-56abd7db-adb4-4bbd-a224-619d17245d67 req-1676c48f-4033-4e9a-8c87-7d1581cf3940 service nova] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Updating instance_info_cache with network_info: [{"id": "42cd0220-15ec-4c55-b1f3-7c583246ff5a", "address": "fa:16:3e:ef:b0:a4", "network": {"id": "884d3248-07f1-4cf7-8ceb-b94b34622384", "bridge": "br-int", "label": "tempest-ServersTestJSON-1882023778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2c6a4814c234d53bbd4844cd482fbae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42cd0220-15", "ovs_interfaceid": "42cd0220-15ec-4c55-b1f3-7c583246ff5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.561175] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250647, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062755} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.561463] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1037.562331] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad74cfed-83e1-4f64-b848-bbe3564a33c6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.586398] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232/f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1037.586732] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dad552fc-f4e2-40fc-9217-554ad6a76e6d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.607473] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1037.607473] env[62569]: value = "task-1250648" [ 1037.607473] env[62569]: _type = "Task" [ 1037.607473] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.615611] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250648, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.695823] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250646, 'name': CreateVM_Task, 'duration_secs': 0.513104} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.696101] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1037.696695] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.696864] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.697215] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1037.697482] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd0582e3-26f6-4162-96de-f048e47b43cc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.705616] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1037.705616] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]521e0b66-09c1-a4be-4d0d-f26669b0bd18" [ 1037.705616] env[62569]: _type = "Task" [ 1037.705616] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.715174] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]521e0b66-09c1-a4be-4d0d-f26669b0bd18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.732548] env[62569]: DEBUG oslo_concurrency.lockutils [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "refresh_cache-96167629-8777-4a1c-b564-c70d1bb59b90" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.732726] env[62569]: DEBUG oslo_concurrency.lockutils [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "refresh_cache-96167629-8777-4a1c-b564-c70d1bb59b90" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.732857] env[62569]: DEBUG nova.network.neutron [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1037.800735] env[62569]: DEBUG nova.scheduler.client.report [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1037.969725] env[62569]: DEBUG oslo_concurrency.lockutils [req-56abd7db-adb4-4bbd-a224-619d17245d67 req-1676c48f-4033-4e9a-8c87-7d1581cf3940 service nova] Releasing lock "refresh_cache-6e3b1490-830d-4869-b805-b83ef0cf9cc7" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.969858] env[62569]: DEBUG nova.compute.manager [req-56abd7db-adb4-4bbd-a224-619d17245d67 req-1676c48f-4033-4e9a-8c87-7d1581cf3940 service nova] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Received event network-vif-deleted-2e8ed1b5-2a5a-4084-b227-177a876c7cf0 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1038.117473] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250648, 'name': ReconfigVM_Task, 'duration_secs': 0.277725} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.117808] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Reconfigured VM instance instance-0000006a to attach disk [datastore1] f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232/f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1038.118431] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fd06f8d7-2406-471e-b948-4c9d70d35606 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.124463] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1038.124463] env[62569]: value = "task-1250649" [ 1038.124463] env[62569]: _type = "Task" [ 1038.124463] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.131655] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250649, 'name': Rename_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.202322] env[62569]: DEBUG nova.network.neutron [-] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.216443] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]521e0b66-09c1-a4be-4d0d-f26669b0bd18, 'name': SearchDatastore_Task, 'duration_secs': 0.008996} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.216885] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.217138] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1038.217693] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.217775] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.217917] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1038.218806] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d5cc4cce-fb65-478c-b65f-f73b0ec05f3b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.228265] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1038.228384] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1038.229153] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-906e13a8-0fb7-49ae-8e24-616560922256 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.237759] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1038.237759] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52489c7d-c3f2-2699-bfbf-cee8bb51d553" [ 1038.237759] env[62569]: _type = "Task" [ 1038.237759] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.247655] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52489c7d-c3f2-2699-bfbf-cee8bb51d553, 'name': SearchDatastore_Task, 'duration_secs': 0.00808} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.248580] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b789d9b-a384-42a4-bd46-53364eb4b512 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.254490] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1038.254490] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520fa4e2-6f1c-d6d0-b570-1c2a812dce73" [ 1038.254490] env[62569]: _type = "Task" [ 1038.254490] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.262264] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520fa4e2-6f1c-d6d0-b570-1c2a812dce73, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.273614] env[62569]: DEBUG nova.network.neutron [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1038.305471] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.717s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.326923] env[62569]: INFO nova.scheduler.client.report [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Deleted allocations for instance 0be076c2-176f-47a9-9489-2f7dfe6d1ce3 [ 1038.412694] env[62569]: DEBUG nova.network.neutron [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Updating instance_info_cache with network_info: [{"id": "d5f53bf1-303b-435e-941a-47c7ab293484", "address": "fa:16:3e:e8:8d:ca", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5f53bf1-30", "ovs_interfaceid": "d5f53bf1-303b-435e-941a-47c7ab293484", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.635441] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250649, 'name': Rename_Task, 'duration_secs': 0.134953} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.635441] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1038.635870] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4eace482-6042-4fc4-8be6-d996f83996b2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.643394] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1038.643394] env[62569]: value = "task-1250650" [ 1038.643394] env[62569]: _type = "Task" [ 1038.643394] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.653494] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250650, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.708447] env[62569]: INFO nova.compute.manager [-] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Took 1.71 seconds to deallocate network for instance. [ 1038.749122] env[62569]: DEBUG nova.compute.manager [req-38bc61b3-a71e-4a9e-a629-07eab3809cb8 req-a0581858-6758-403e-a3c3-d1d2823f6901 service nova] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Received event network-vif-plugged-d5f53bf1-303b-435e-941a-47c7ab293484 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1038.749430] env[62569]: DEBUG oslo_concurrency.lockutils [req-38bc61b3-a71e-4a9e-a629-07eab3809cb8 req-a0581858-6758-403e-a3c3-d1d2823f6901 service nova] Acquiring lock "96167629-8777-4a1c-b564-c70d1bb59b90-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.749720] env[62569]: DEBUG oslo_concurrency.lockutils [req-38bc61b3-a71e-4a9e-a629-07eab3809cb8 req-a0581858-6758-403e-a3c3-d1d2823f6901 service nova] Lock "96167629-8777-4a1c-b564-c70d1bb59b90-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.749975] env[62569]: DEBUG oslo_concurrency.lockutils [req-38bc61b3-a71e-4a9e-a629-07eab3809cb8 req-a0581858-6758-403e-a3c3-d1d2823f6901 service nova] Lock "96167629-8777-4a1c-b564-c70d1bb59b90-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.750217] env[62569]: DEBUG nova.compute.manager [req-38bc61b3-a71e-4a9e-a629-07eab3809cb8 req-a0581858-6758-403e-a3c3-d1d2823f6901 service nova] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] No waiting events found dispatching network-vif-plugged-d5f53bf1-303b-435e-941a-47c7ab293484 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1038.750460] env[62569]: WARNING nova.compute.manager [req-38bc61b3-a71e-4a9e-a629-07eab3809cb8 req-a0581858-6758-403e-a3c3-d1d2823f6901 service nova] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Received unexpected event network-vif-plugged-d5f53bf1-303b-435e-941a-47c7ab293484 for instance with vm_state building and task_state spawning. [ 1038.750700] env[62569]: DEBUG nova.compute.manager [req-38bc61b3-a71e-4a9e-a629-07eab3809cb8 req-a0581858-6758-403e-a3c3-d1d2823f6901 service nova] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Received event network-changed-d5f53bf1-303b-435e-941a-47c7ab293484 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1038.750925] env[62569]: DEBUG nova.compute.manager [req-38bc61b3-a71e-4a9e-a629-07eab3809cb8 req-a0581858-6758-403e-a3c3-d1d2823f6901 service nova] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Refreshing instance network info cache due to event network-changed-d5f53bf1-303b-435e-941a-47c7ab293484. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1038.751207] env[62569]: DEBUG oslo_concurrency.lockutils [req-38bc61b3-a71e-4a9e-a629-07eab3809cb8 req-a0581858-6758-403e-a3c3-d1d2823f6901 service nova] Acquiring lock "refresh_cache-96167629-8777-4a1c-b564-c70d1bb59b90" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.764673] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]520fa4e2-6f1c-d6d0-b570-1c2a812dce73, 'name': SearchDatastore_Task, 'duration_secs': 0.007525} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.764922] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.765179] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 6e3b1490-830d-4869-b805-b83ef0cf9cc7/6e3b1490-830d-4869-b805-b83ef0cf9cc7.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1038.765430] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3fdc9c88-7e7c-4fa0-9a78-4acf546a85a8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.772024] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1038.772024] env[62569]: value = "task-1250651" [ 1038.772024] env[62569]: _type = "Task" [ 1038.772024] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.779540] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250651, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.835976] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7e24c360-2429-4aa1-b12d-f5e9abe8a199 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Lock "0be076c2-176f-47a9-9489-2f7dfe6d1ce3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.694s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.915345] env[62569]: DEBUG oslo_concurrency.lockutils [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "refresh_cache-96167629-8777-4a1c-b564-c70d1bb59b90" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.915685] env[62569]: DEBUG nova.compute.manager [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Instance network_info: |[{"id": "d5f53bf1-303b-435e-941a-47c7ab293484", "address": "fa:16:3e:e8:8d:ca", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5f53bf1-30", "ovs_interfaceid": "d5f53bf1-303b-435e-941a-47c7ab293484", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1038.916027] env[62569]: DEBUG oslo_concurrency.lockutils [req-38bc61b3-a71e-4a9e-a629-07eab3809cb8 req-a0581858-6758-403e-a3c3-d1d2823f6901 service nova] Acquired lock "refresh_cache-96167629-8777-4a1c-b564-c70d1bb59b90" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.916438] env[62569]: DEBUG nova.network.neutron [req-38bc61b3-a71e-4a9e-a629-07eab3809cb8 req-a0581858-6758-403e-a3c3-d1d2823f6901 service nova] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Refreshing network info cache for port d5f53bf1-303b-435e-941a-47c7ab293484 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1038.917870] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e8:8d:ca', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73915082-a1b0-460b-b24d-97588fc9cb29', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd5f53bf1-303b-435e-941a-47c7ab293484', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1038.925893] env[62569]: DEBUG oslo.service.loopingcall [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1038.928921] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1038.929556] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1865c41b-da89-4f38-b4e8-091131566d86 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.955338] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1038.955338] env[62569]: value = "task-1250652" [ 1038.955338] env[62569]: _type = "Task" [ 1038.955338] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.963955] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250652, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.156894] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250650, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.177601] env[62569]: DEBUG nova.network.neutron [req-38bc61b3-a71e-4a9e-a629-07eab3809cb8 req-a0581858-6758-403e-a3c3-d1d2823f6901 service nova] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Updated VIF entry in instance network info cache for port d5f53bf1-303b-435e-941a-47c7ab293484. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1039.178099] env[62569]: DEBUG nova.network.neutron [req-38bc61b3-a71e-4a9e-a629-07eab3809cb8 req-a0581858-6758-403e-a3c3-d1d2823f6901 service nova] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Updating instance_info_cache with network_info: [{"id": "d5f53bf1-303b-435e-941a-47c7ab293484", "address": "fa:16:3e:e8:8d:ca", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5f53bf1-30", "ovs_interfaceid": "d5f53bf1-303b-435e-941a-47c7ab293484", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.214324] env[62569]: DEBUG oslo_concurrency.lockutils [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.214643] env[62569]: DEBUG oslo_concurrency.lockutils [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.214877] env[62569]: DEBUG nova.objects.instance [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lazy-loading 'resources' on Instance uuid d56faf39-6caf-4489-98d7-342fc5fb7d40 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1039.285276] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250651, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.452521} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.285653] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 6e3b1490-830d-4869-b805-b83ef0cf9cc7/6e3b1490-830d-4869-b805-b83ef0cf9cc7.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1039.286359] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1039.286359] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-91f22075-54b6-49ab-8da4-4ce2685274ce {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.293092] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1039.293092] env[62569]: value = "task-1250653" [ 1039.293092] env[62569]: _type = "Task" [ 1039.293092] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.301026] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250653, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.464727] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250652, 'name': CreateVM_Task, 'duration_secs': 0.42588} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.464910] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1039.465651] env[62569]: DEBUG oslo_concurrency.lockutils [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.465778] env[62569]: DEBUG oslo_concurrency.lockutils [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.466127] env[62569]: DEBUG oslo_concurrency.lockutils [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1039.466384] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2d35025-b93b-44a1-b3e3-e3c32b66525a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.471061] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1039.471061] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52599139-a1b6-ee18-1686-30824d029de3" [ 1039.471061] env[62569]: _type = "Task" [ 1039.471061] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.478660] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52599139-a1b6-ee18-1686-30824d029de3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.654205] env[62569]: DEBUG oslo_vmware.api [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250650, 'name': PowerOnVM_Task, 'duration_secs': 0.55596} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.654512] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1039.654716] env[62569]: INFO nova.compute.manager [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Took 7.21 seconds to spawn the instance on the hypervisor. [ 1039.654897] env[62569]: DEBUG nova.compute.manager [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1039.655659] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0256523f-bce5-4ce8-bacf-f2dfecf0298f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.681296] env[62569]: DEBUG oslo_concurrency.lockutils [req-38bc61b3-a71e-4a9e-a629-07eab3809cb8 req-a0581858-6758-403e-a3c3-d1d2823f6901 service nova] Releasing lock "refresh_cache-96167629-8777-4a1c-b564-c70d1bb59b90" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.681535] env[62569]: DEBUG nova.compute.manager [req-38bc61b3-a71e-4a9e-a629-07eab3809cb8 req-a0581858-6758-403e-a3c3-d1d2823f6901 service nova] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Received event network-vif-deleted-b6625a0e-7282-4ade-b7f0-3e5374ec8e64 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1039.804093] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250653, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064766} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.804395] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1039.805229] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b1baf3a-ef4b-4b5c-a17d-fb66da04ac0e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.809040] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d15ff962-6314-474a-adc0-4e4cbe1fa161 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.824132] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d624c103-5e16-4cf1-adeb-54205631aebf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.835456] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 6e3b1490-830d-4869-b805-b83ef0cf9cc7/6e3b1490-830d-4869-b805-b83ef0cf9cc7.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1039.836043] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9c15aea-3766-4b8b-b561-3256dc2b9b3d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.878331] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquiring lock "9b5a1451-af46-4cd8-8e7a-f4560d8499e6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.878599] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Lock "9b5a1451-af46-4cd8-8e7a-f4560d8499e6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.878793] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquiring lock "9b5a1451-af46-4cd8-8e7a-f4560d8499e6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.878976] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Lock "9b5a1451-af46-4cd8-8e7a-f4560d8499e6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.879490] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Lock "9b5a1451-af46-4cd8-8e7a-f4560d8499e6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.881660] env[62569]: INFO nova.compute.manager [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Terminating instance [ 1039.886447] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6df43fc-5b77-4466-a580-3a9ce11ccd01 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.889234] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1039.889234] env[62569]: value = "task-1250654" [ 1039.889234] env[62569]: _type = "Task" [ 1039.889234] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.897202] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5432c7dc-0a8e-404e-bf9f-1b9e1f0a0ea8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.904744] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250654, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.915204] env[62569]: DEBUG nova.compute.provider_tree [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.981324] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52599139-a1b6-ee18-1686-30824d029de3, 'name': SearchDatastore_Task, 'duration_secs': 0.008644} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.981656] env[62569]: DEBUG oslo_concurrency.lockutils [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.981926] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1039.982225] env[62569]: DEBUG oslo_concurrency.lockutils [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.982407] env[62569]: DEBUG oslo_concurrency.lockutils [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.982624] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1039.982916] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9e756e0c-431b-428b-b2a1-1f64179dfd9a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.992516] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1039.992920] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1039.993705] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-619e82de-6dcd-4a8e-808a-b191b611b51e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.998733] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1039.998733] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d9b490-a915-a0be-97da-9556f6f7d2a4" [ 1039.998733] env[62569]: _type = "Task" [ 1039.998733] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.006580] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d9b490-a915-a0be-97da-9556f6f7d2a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.172630] env[62569]: INFO nova.compute.manager [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Took 11.94 seconds to build instance. [ 1040.385869] env[62569]: DEBUG nova.compute.manager [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1040.386139] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1040.387064] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e161b6ab-20e8-409d-91ab-8affb12cdc40 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.395390] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1040.395937] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5ed465b6-e916-458e-a2c3-23a4748bb4c9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.400410] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250654, 'name': ReconfigVM_Task, 'duration_secs': 0.302529} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.401016] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 6e3b1490-830d-4869-b805-b83ef0cf9cc7/6e3b1490-830d-4869-b805-b83ef0cf9cc7.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1040.401640] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f3d43ee8-d905-4408-ab2f-f1c2d307a9b5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.405642] env[62569]: DEBUG oslo_vmware.api [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 1040.405642] env[62569]: value = "task-1250655" [ 1040.405642] env[62569]: _type = "Task" [ 1040.405642] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.409484] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1040.409484] env[62569]: value = "task-1250656" [ 1040.409484] env[62569]: _type = "Task" [ 1040.409484] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.417871] env[62569]: DEBUG nova.scheduler.client.report [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1040.420960] env[62569]: DEBUG oslo_vmware.api [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250655, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.427261] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250656, 'name': Rename_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.508890] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52d9b490-a915-a0be-97da-9556f6f7d2a4, 'name': SearchDatastore_Task, 'duration_secs': 0.011712} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.509711] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5fda15d2-ed2b-4059-ba2b-50db29a4ffcd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.515606] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1040.515606] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52755e3c-0d04-8bf1-1121-4fcfde7ec041" [ 1040.515606] env[62569]: _type = "Task" [ 1040.515606] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.523291] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52755e3c-0d04-8bf1-1121-4fcfde7ec041, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.674962] env[62569]: DEBUG oslo_concurrency.lockutils [None req-3179b725-7d1a-414f-a742-29b9d9a79298 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.452s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.916881] env[62569]: DEBUG oslo_vmware.api [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250655, 'name': PowerOffVM_Task, 'duration_secs': 0.219086} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.917534] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1040.917735] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1040.917976] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-43ab1cb9-b623-43a7-ab44-0d35026d1a51 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.921857] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250656, 'name': Rename_Task, 'duration_secs': 0.16887} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.922812] env[62569]: DEBUG oslo_concurrency.lockutils [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.708s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.924589] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1040.925017] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8f75fee9-1f37-46ea-b628-8d58a42217fe {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.930650] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1040.930650] env[62569]: value = "task-1250658" [ 1040.930650] env[62569]: _type = "Task" [ 1040.930650] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.938216] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250658, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.942717] env[62569]: INFO nova.scheduler.client.report [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Deleted allocations for instance d56faf39-6caf-4489-98d7-342fc5fb7d40 [ 1040.985194] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1040.985443] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1040.985636] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Deleting the datastore file [datastore1] 9b5a1451-af46-4cd8-8e7a-f4560d8499e6 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1040.985917] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5e61d77a-acb0-4679-96a9-84ca5ced9b0a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.992577] env[62569]: DEBUG oslo_vmware.api [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for the task: (returnval){ [ 1040.992577] env[62569]: value = "task-1250659" [ 1040.992577] env[62569]: _type = "Task" [ 1040.992577] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.000200] env[62569]: DEBUG oslo_vmware.api [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250659, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.025268] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52755e3c-0d04-8bf1-1121-4fcfde7ec041, 'name': SearchDatastore_Task, 'duration_secs': 0.010493} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.025525] env[62569]: DEBUG oslo_concurrency.lockutils [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.025783] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 96167629-8777-4a1c-b564-c70d1bb59b90/96167629-8777-4a1c-b564-c70d1bb59b90.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1041.026047] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-51b6b293-a0db-410e-87d4-9e2b061bd24a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.031501] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1041.031501] env[62569]: value = "task-1250660" [ 1041.031501] env[62569]: _type = "Task" [ 1041.031501] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.038710] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250660, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.124614] env[62569]: DEBUG nova.compute.manager [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Stashing vm_state: active {{(pid=62569) _prep_resize /opt/stack/nova/nova/compute/manager.py:5997}} [ 1041.444142] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250658, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.449709] env[62569]: DEBUG oslo_concurrency.lockutils [None req-40b7e838-a884-4692-aef5-822c1d54971f tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "d56faf39-6caf-4489-98d7-342fc5fb7d40" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.066s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.505572] env[62569]: DEBUG oslo_vmware.api [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Task: {'id': task-1250659, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.226899} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.505905] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1041.506130] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1041.506323] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1041.506501] env[62569]: INFO nova.compute.manager [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1041.506747] env[62569]: DEBUG oslo.service.loopingcall [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1041.506947] env[62569]: DEBUG nova.compute.manager [-] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1041.507082] env[62569]: DEBUG nova.network.neutron [-] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1041.543083] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250660, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.479027} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.543083] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore2] 96167629-8777-4a1c-b564-c70d1bb59b90/96167629-8777-4a1c-b564-c70d1bb59b90.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1041.543083] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1041.543421] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-33aa5741-e8fd-4165-8584-9d3a27431060 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.550158] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1041.550158] env[62569]: value = "task-1250661" [ 1041.550158] env[62569]: _type = "Task" [ 1041.550158] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.556737] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250661, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.646118] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.646299] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.803741] env[62569]: DEBUG nova.compute.manager [req-699e3d7e-eed7-40d8-ba6a-617a3368e2ed req-7ab7f712-f542-445e-9aa1-4d8a73a012f8 service nova] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Received event network-vif-deleted-2e9fac89-877b-4c36-808d-21c6f01c992d {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1041.803983] env[62569]: INFO nova.compute.manager [req-699e3d7e-eed7-40d8-ba6a-617a3368e2ed req-7ab7f712-f542-445e-9aa1-4d8a73a012f8 service nova] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Neutron deleted interface 2e9fac89-877b-4c36-808d-21c6f01c992d; detaching it from the instance and deleting it from the info cache [ 1041.804154] env[62569]: DEBUG nova.network.neutron [req-699e3d7e-eed7-40d8-ba6a-617a3368e2ed req-7ab7f712-f542-445e-9aa1-4d8a73a012f8 service nova] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.942447] env[62569]: DEBUG oslo_vmware.api [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250658, 'name': PowerOnVM_Task, 'duration_secs': 0.613084} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.942745] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1041.942944] env[62569]: INFO nova.compute.manager [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Took 7.31 seconds to spawn the instance on the hypervisor. [ 1041.943145] env[62569]: DEBUG nova.compute.manager [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1041.943956] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-417a0836-3d98-4f90-bccd-c74d4ad061b0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.060440] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250661, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085095} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.061186] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1042.062386] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c389f64-59a2-4885-8097-199a189d4468 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.087273] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] 96167629-8777-4a1c-b564-c70d1bb59b90/96167629-8777-4a1c-b564-c70d1bb59b90.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1042.087596] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6fcc382c-8bc4-4651-a78d-76ebdddf29e3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.106979] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1042.106979] env[62569]: value = "task-1250662" [ 1042.106979] env[62569]: _type = "Task" [ 1042.106979] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.115015] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250662, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.151357] env[62569]: INFO nova.compute.claims [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1042.280555] env[62569]: DEBUG nova.network.neutron [-] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.307311] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a2db664c-33a9-47e3-9589-884b52bd1bcb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.318800] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-970f0a2e-604c-4364-bb7c-952a91a7157e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.346515] env[62569]: DEBUG nova.compute.manager [req-699e3d7e-eed7-40d8-ba6a-617a3368e2ed req-7ab7f712-f542-445e-9aa1-4d8a73a012f8 service nova] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Detach interface failed, port_id=2e9fac89-877b-4c36-808d-21c6f01c992d, reason: Instance 9b5a1451-af46-4cd8-8e7a-f4560d8499e6 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1042.461960] env[62569]: INFO nova.compute.manager [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Took 12.80 seconds to build instance. [ 1042.616184] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250662, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.658088] env[62569]: INFO nova.compute.resource_tracker [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Updating resource usage from migration de882a97-fdf6-4238-9efb-f038b01e1cf6 [ 1042.767686] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fd07321-bf02-4a44-b9e7-484fae769545 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.775782] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdc9f825-82e5-4381-a2fb-e076cb08886e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.804540] env[62569]: INFO nova.compute.manager [-] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Took 1.30 seconds to deallocate network for instance. [ 1042.807057] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9920c99e-abcc-415f-b36e-9512c0a84d1f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.817203] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6736e4a1-1c6a-4083-aa1b-bb39f7759314 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.831901] env[62569]: DEBUG nova.compute.provider_tree [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1042.964421] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47ebe37a-d83c-4f69-a99a-24940a3ddefa tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "6e3b1490-830d-4869-b805-b83ef0cf9cc7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.309s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.116978] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250662, 'name': ReconfigVM_Task, 'duration_secs': 0.981432} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.117382] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Reconfigured VM instance instance-0000006c to attach disk [datastore2] 96167629-8777-4a1c-b564-c70d1bb59b90/96167629-8777-4a1c-b564-c70d1bb59b90.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1043.117999] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7b671beb-e438-4e54-a96a-96a793d1eafc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.125199] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1043.125199] env[62569]: value = "task-1250663" [ 1043.125199] env[62569]: _type = "Task" [ 1043.125199] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.137194] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250663, 'name': Rename_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.314664] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.335867] env[62569]: DEBUG nova.scheduler.client.report [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1043.472475] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "6e3b1490-830d-4869-b805-b83ef0cf9cc7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.472752] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "6e3b1490-830d-4869-b805-b83ef0cf9cc7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.472968] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "6e3b1490-830d-4869-b805-b83ef0cf9cc7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.473220] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "6e3b1490-830d-4869-b805-b83ef0cf9cc7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.473405] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "6e3b1490-830d-4869-b805-b83ef0cf9cc7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.476007] env[62569]: INFO nova.compute.manager [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Terminating instance [ 1043.635270] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250663, 'name': Rename_Task, 'duration_secs': 0.286578} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.635549] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1043.635795] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2a0e0cc4-67d2-46b3-9ba2-24246ba3657b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.641978] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1043.641978] env[62569]: value = "task-1250664" [ 1043.641978] env[62569]: _type = "Task" [ 1043.641978] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.652901] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250664, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.840938] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.195s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.841195] env[62569]: INFO nova.compute.manager [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Migrating [ 1043.848740] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.534s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.848979] env[62569]: DEBUG nova.objects.instance [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Lazy-loading 'resources' on Instance uuid 9b5a1451-af46-4cd8-8e7a-f4560d8499e6 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.860224] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "cf077eb8-8860-46f0-ae84-eed58c29873b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.860461] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cf077eb8-8860-46f0-ae84-eed58c29873b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.981062] env[62569]: DEBUG nova.compute.manager [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1043.981062] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1043.981677] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a831a06d-238c-4e97-809e-478d68f6f6e3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.989897] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1043.990177] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b5f78d3d-1315-4be1-87bb-deec8f3bd43f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.997027] env[62569]: DEBUG oslo_vmware.api [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1043.997027] env[62569]: value = "task-1250665" [ 1043.997027] env[62569]: _type = "Task" [ 1043.997027] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.007243] env[62569]: DEBUG oslo_vmware.api [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250665, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.152338] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250664, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.361389] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "refresh_cache-f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.361622] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired lock "refresh_cache-f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.361806] env[62569]: DEBUG nova.network.neutron [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1044.366355] env[62569]: DEBUG nova.compute.manager [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1044.465885] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4576f9f6-6d71-4da9-850c-868319c800ae {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.473479] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e65973c7-46a8-42da-b344-da55c2e5d733 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.504534] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73f442f4-4861-4258-bfbc-68ad6b909411 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.514039] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddb12981-8326-4cc0-a8d6-920cde59fc51 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.517529] env[62569]: DEBUG oslo_vmware.api [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250665, 'name': PowerOffVM_Task, 'duration_secs': 0.194575} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.517776] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1044.517946] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1044.518514] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-73193ac4-e897-4859-9502-d4f33524b27c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.527482] env[62569]: DEBUG nova.compute.provider_tree [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1044.579864] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1044.580324] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1044.580630] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Deleting the datastore file [datastore1] 6e3b1490-830d-4869-b805-b83ef0cf9cc7 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1044.580958] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f10dd5dc-c654-4d30-9877-c1ca2fadc252 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.587385] env[62569]: DEBUG oslo_vmware.api [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1044.587385] env[62569]: value = "task-1250667" [ 1044.587385] env[62569]: _type = "Task" [ 1044.587385] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.596462] env[62569]: DEBUG oslo_vmware.api [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250667, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.652913] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250664, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.886573] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.030560] env[62569]: DEBUG nova.scheduler.client.report [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1045.055468] env[62569]: DEBUG nova.network.neutron [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Updating instance_info_cache with network_info: [{"id": "04dded5b-c41d-4dd5-b9bb-354115b85a5d", "address": "fa:16:3e:50:e2:7f", "network": {"id": "30d3bf4a-8f55-492f-a049-181e29731818", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1244949475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bdba6022e3d4697a336ca28ca4eccec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04dded5b-c4", "ovs_interfaceid": "04dded5b-c41d-4dd5-b9bb-354115b85a5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.097217] env[62569]: DEBUG oslo_vmware.api [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250667, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177093} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.097721] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1045.097914] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1045.098104] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1045.098277] env[62569]: INFO nova.compute.manager [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1045.098515] env[62569]: DEBUG oslo.service.loopingcall [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1045.098701] env[62569]: DEBUG nova.compute.manager [-] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1045.098793] env[62569]: DEBUG nova.network.neutron [-] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1045.153073] env[62569]: DEBUG oslo_vmware.api [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250664, 'name': PowerOnVM_Task, 'duration_secs': 1.208688} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.153305] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1045.153526] env[62569]: INFO nova.compute.manager [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Took 8.32 seconds to spawn the instance on the hypervisor. [ 1045.153714] env[62569]: DEBUG nova.compute.manager [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1045.154480] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c23728d-5bc0-45d8-b9d3-aded303ded7f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.359386] env[62569]: DEBUG nova.compute.manager [req-179e7938-2c98-4fcf-927c-31746d12fdd9 req-5f9a6a55-2067-4512-95d4-76d0947a9f3d service nova] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Received event network-vif-deleted-42cd0220-15ec-4c55-b1f3-7c583246ff5a {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1045.359386] env[62569]: INFO nova.compute.manager [req-179e7938-2c98-4fcf-927c-31746d12fdd9 req-5f9a6a55-2067-4512-95d4-76d0947a9f3d service nova] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Neutron deleted interface 42cd0220-15ec-4c55-b1f3-7c583246ff5a; detaching it from the instance and deleting it from the info cache [ 1045.359386] env[62569]: DEBUG nova.network.neutron [req-179e7938-2c98-4fcf-927c-31746d12fdd9 req-5f9a6a55-2067-4512-95d4-76d0947a9f3d service nova] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.520888] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c244e362-a91f-4ea1-882a-472f1b2f7161 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.521225] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c244e362-a91f-4ea1-882a-472f1b2f7161 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.535618] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.687s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.537821] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.651s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.539334] env[62569]: INFO nova.compute.claims [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1045.557529] env[62569]: INFO nova.scheduler.client.report [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Deleted allocations for instance 9b5a1451-af46-4cd8-8e7a-f4560d8499e6 [ 1045.558548] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Releasing lock "refresh_cache-f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.670569] env[62569]: INFO nova.compute.manager [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Took 15.09 seconds to build instance. [ 1045.840408] env[62569]: DEBUG nova.network.neutron [-] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.862339] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e130fe3f-5556-432c-a460-92efccf6e3ca {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.873984] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab7653a8-7e28-420b-a7a0-7d9b74702eac {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.899066] env[62569]: DEBUG nova.compute.manager [req-179e7938-2c98-4fcf-927c-31746d12fdd9 req-5f9a6a55-2067-4512-95d4-76d0947a9f3d service nova] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Detach interface failed, port_id=42cd0220-15ec-4c55-b1f3-7c583246ff5a, reason: Instance 6e3b1490-830d-4869-b805-b83ef0cf9cc7 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1046.024977] env[62569]: DEBUG nova.compute.utils [None req-c244e362-a91f-4ea1-882a-472f1b2f7161 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1046.069535] env[62569]: DEBUG oslo_concurrency.lockutils [None req-1d36fa3f-8535-4dca-9511-6148c73b6f71 tempest-ServerRescueTestJSON-1535232041 tempest-ServerRescueTestJSON-1535232041-project-member] Lock "9b5a1451-af46-4cd8-8e7a-f4560d8499e6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.191s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.173025] env[62569]: DEBUG oslo_concurrency.lockutils [None req-90d98c68-b85c-4d92-87bc-f45ce5f14504 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "96167629-8777-4a1c-b564-c70d1bb59b90" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.597s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.343321] env[62569]: INFO nova.compute.manager [-] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Took 1.24 seconds to deallocate network for instance. [ 1046.421801] env[62569]: DEBUG nova.compute.manager [req-bd84654b-0788-421c-af93-13e0ddf37df9 req-daf3ef2b-312d-4b68-8899-d565ce3e196b service nova] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Received event network-changed-d5f53bf1-303b-435e-941a-47c7ab293484 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1046.421801] env[62569]: DEBUG nova.compute.manager [req-bd84654b-0788-421c-af93-13e0ddf37df9 req-daf3ef2b-312d-4b68-8899-d565ce3e196b service nova] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Refreshing instance network info cache due to event network-changed-d5f53bf1-303b-435e-941a-47c7ab293484. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1046.421801] env[62569]: DEBUG oslo_concurrency.lockutils [req-bd84654b-0788-421c-af93-13e0ddf37df9 req-daf3ef2b-312d-4b68-8899-d565ce3e196b service nova] Acquiring lock "refresh_cache-96167629-8777-4a1c-b564-c70d1bb59b90" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.421801] env[62569]: DEBUG oslo_concurrency.lockutils [req-bd84654b-0788-421c-af93-13e0ddf37df9 req-daf3ef2b-312d-4b68-8899-d565ce3e196b service nova] Acquired lock "refresh_cache-96167629-8777-4a1c-b564-c70d1bb59b90" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.422214] env[62569]: DEBUG nova.network.neutron [req-bd84654b-0788-421c-af93-13e0ddf37df9 req-daf3ef2b-312d-4b68-8899-d565ce3e196b service nova] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Refreshing network info cache for port d5f53bf1-303b-435e-941a-47c7ab293484 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1046.528653] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c244e362-a91f-4ea1-882a-472f1b2f7161 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.658031] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f8262d5-77a0-4b78-a8dc-e963c1382030 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.666592] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32559b4c-ee65-4455-aceb-a2c4ddcb5bce {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.699150] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff403b76-1e2b-4425-8271-23bbd6a2db5f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.706720] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2257f0f-1b2e-46da-8a29-1a3b9dc1cdee {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.720095] env[62569]: DEBUG nova.compute.provider_tree [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1046.850317] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.085114] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29c0b849-108c-4f4b-be9f-c57b74daeabe {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.104585] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Updating instance 'f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232' progress to 0 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1047.225993] env[62569]: DEBUG nova.scheduler.client.report [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1047.271296] env[62569]: DEBUG nova.network.neutron [req-bd84654b-0788-421c-af93-13e0ddf37df9 req-daf3ef2b-312d-4b68-8899-d565ce3e196b service nova] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Updated VIF entry in instance network info cache for port d5f53bf1-303b-435e-941a-47c7ab293484. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1047.271296] env[62569]: DEBUG nova.network.neutron [req-bd84654b-0788-421c-af93-13e0ddf37df9 req-daf3ef2b-312d-4b68-8899-d565ce3e196b service nova] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Updating instance_info_cache with network_info: [{"id": "d5f53bf1-303b-435e-941a-47c7ab293484", "address": "fa:16:3e:e8:8d:ca", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5f53bf1-30", "ovs_interfaceid": "d5f53bf1-303b-435e-941a-47c7ab293484", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.612733] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1047.613135] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0706ac63-368e-4620-bf8c-bd431a883c6d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.620163] env[62569]: DEBUG oslo_vmware.api [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1047.620163] env[62569]: value = "task-1250668" [ 1047.620163] env[62569]: _type = "Task" [ 1047.620163] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.627857] env[62569]: DEBUG oslo_vmware.api [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250668, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.735204] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.197s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.735709] env[62569]: DEBUG nova.compute.manager [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1047.738690] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.889s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.738923] env[62569]: DEBUG nova.objects.instance [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lazy-loading 'resources' on Instance uuid 6e3b1490-830d-4869-b805-b83ef0cf9cc7 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1047.774091] env[62569]: DEBUG oslo_concurrency.lockutils [req-bd84654b-0788-421c-af93-13e0ddf37df9 req-daf3ef2b-312d-4b68-8899-d565ce3e196b service nova] Releasing lock "refresh_cache-96167629-8777-4a1c-b564-c70d1bb59b90" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.799149] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c244e362-a91f-4ea1-882a-472f1b2f7161 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.799495] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c244e362-a91f-4ea1-882a-472f1b2f7161 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.799737] env[62569]: INFO nova.compute.manager [None req-c244e362-a91f-4ea1-882a-472f1b2f7161 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Attaching volume aa9283c3-c013-498f-8986-2367724e045f to /dev/sdb [ 1047.832189] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dc3810e-e667-4390-904c-6723302c7fb2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.838845] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37447dbf-0b30-4ea0-8497-80af28884168 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.854213] env[62569]: DEBUG nova.virt.block_device [None req-c244e362-a91f-4ea1-882a-472f1b2f7161 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Updating existing volume attachment record: 06373b9c-c170-4e20-b126-543b4203c8e4 {{(pid=62569) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1048.132114] env[62569]: DEBUG oslo_vmware.api [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250668, 'name': PowerOffVM_Task, 'duration_secs': 0.17699} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.132569] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1048.132892] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Updating instance 'f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232' progress to 17 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1048.241901] env[62569]: DEBUG nova.compute.utils [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1048.246297] env[62569]: DEBUG nova.compute.manager [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1048.246501] env[62569]: DEBUG nova.network.neutron [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1048.295676] env[62569]: DEBUG nova.policy [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '50780a0da3b14122aa6287c1a482f671', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7f7fb3e698324f8e851aefbee74ac669', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 1048.348455] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68d79197-f957-48a7-940c-7ed55f79f703 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.355884] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9ac50d0-c16a-4138-bc12-4497d47202e7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.386411] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a54cb3e-84c6-40f7-9f43-23f6d785b824 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.394625] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95370dac-de33-4418-b7f3-dcf5665f9bcf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.407766] env[62569]: DEBUG nova.compute.provider_tree [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1048.581464] env[62569]: DEBUG nova.network.neutron [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Successfully created port: 57d52bb3-5c0a-4ea0-89b3-94d652923656 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1048.641770] env[62569]: DEBUG nova.virt.hardware [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1048.643612] env[62569]: DEBUG nova.virt.hardware [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1048.644321] env[62569]: DEBUG nova.virt.hardware [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1048.645060] env[62569]: DEBUG nova.virt.hardware [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1048.645574] env[62569]: DEBUG nova.virt.hardware [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1048.645784] env[62569]: DEBUG nova.virt.hardware [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1048.646330] env[62569]: DEBUG nova.virt.hardware [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1048.647083] env[62569]: DEBUG nova.virt.hardware [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1048.647336] env[62569]: DEBUG nova.virt.hardware [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1048.647545] env[62569]: DEBUG nova.virt.hardware [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1048.647760] env[62569]: DEBUG nova.virt.hardware [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1048.657189] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de16188c-3b7c-4f9a-ac8e-cdf1581c7112 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.673826] env[62569]: DEBUG oslo_vmware.api [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1048.673826] env[62569]: value = "task-1250672" [ 1048.673826] env[62569]: _type = "Task" [ 1048.673826] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.683888] env[62569]: DEBUG oslo_vmware.api [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250672, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.749781] env[62569]: DEBUG nova.compute.manager [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1048.911455] env[62569]: DEBUG nova.scheduler.client.report [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1049.184311] env[62569]: DEBUG oslo_vmware.api [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250672, 'name': ReconfigVM_Task, 'duration_secs': 0.153694} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.184636] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Updating instance 'f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232' progress to 33 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1049.416535] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.678s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.438101] env[62569]: INFO nova.scheduler.client.report [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Deleted allocations for instance 6e3b1490-830d-4869-b805-b83ef0cf9cc7 [ 1049.691366] env[62569]: DEBUG nova.virt.hardware [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1049.691897] env[62569]: DEBUG nova.virt.hardware [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1049.691897] env[62569]: DEBUG nova.virt.hardware [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1049.692036] env[62569]: DEBUG nova.virt.hardware [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1049.692197] env[62569]: DEBUG nova.virt.hardware [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1049.692371] env[62569]: DEBUG nova.virt.hardware [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1049.692590] env[62569]: DEBUG nova.virt.hardware [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1049.692755] env[62569]: DEBUG nova.virt.hardware [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1049.692926] env[62569]: DEBUG nova.virt.hardware [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1049.693106] env[62569]: DEBUG nova.virt.hardware [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1049.693293] env[62569]: DEBUG nova.virt.hardware [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1049.698590] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Reconfiguring VM instance instance-0000006a to detach disk 2000 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1049.698900] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5a825bce-5b85-4693-9b2e-482f6d2f15bf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.718214] env[62569]: DEBUG oslo_vmware.api [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1049.718214] env[62569]: value = "task-1250673" [ 1049.718214] env[62569]: _type = "Task" [ 1049.718214] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.727923] env[62569]: DEBUG oslo_vmware.api [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250673, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.759973] env[62569]: DEBUG nova.compute.manager [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1049.781216] env[62569]: DEBUG nova.virt.hardware [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1049.781479] env[62569]: DEBUG nova.virt.hardware [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1049.781646] env[62569]: DEBUG nova.virt.hardware [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1049.781831] env[62569]: DEBUG nova.virt.hardware [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1049.781984] env[62569]: DEBUG nova.virt.hardware [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1049.782150] env[62569]: DEBUG nova.virt.hardware [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1049.782454] env[62569]: DEBUG nova.virt.hardware [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1049.782642] env[62569]: DEBUG nova.virt.hardware [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1049.782852] env[62569]: DEBUG nova.virt.hardware [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1049.783058] env[62569]: DEBUG nova.virt.hardware [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1049.783305] env[62569]: DEBUG nova.virt.hardware [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1049.784213] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b872c9-2b66-4da4-8119-1d0248189c5c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.792406] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b2e6d87-2d2c-4544-9c57-121bd9495d3a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.942820] env[62569]: DEBUG nova.compute.manager [req-c3ca3134-4fb2-42c9-bf10-6cbb52e4b541 req-622a1aeb-3c5e-4dbc-9b6a-506ea1cdada2 service nova] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Received event network-vif-plugged-57d52bb3-5c0a-4ea0-89b3-94d652923656 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1049.943471] env[62569]: DEBUG oslo_concurrency.lockutils [req-c3ca3134-4fb2-42c9-bf10-6cbb52e4b541 req-622a1aeb-3c5e-4dbc-9b6a-506ea1cdada2 service nova] Acquiring lock "cf077eb8-8860-46f0-ae84-eed58c29873b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.943471] env[62569]: DEBUG oslo_concurrency.lockutils [req-c3ca3134-4fb2-42c9-bf10-6cbb52e4b541 req-622a1aeb-3c5e-4dbc-9b6a-506ea1cdada2 service nova] Lock "cf077eb8-8860-46f0-ae84-eed58c29873b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.943782] env[62569]: DEBUG oslo_concurrency.lockutils [req-c3ca3134-4fb2-42c9-bf10-6cbb52e4b541 req-622a1aeb-3c5e-4dbc-9b6a-506ea1cdada2 service nova] Lock "cf077eb8-8860-46f0-ae84-eed58c29873b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.943905] env[62569]: DEBUG nova.compute.manager [req-c3ca3134-4fb2-42c9-bf10-6cbb52e4b541 req-622a1aeb-3c5e-4dbc-9b6a-506ea1cdada2 service nova] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] No waiting events found dispatching network-vif-plugged-57d52bb3-5c0a-4ea0-89b3-94d652923656 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1049.944140] env[62569]: WARNING nova.compute.manager [req-c3ca3134-4fb2-42c9-bf10-6cbb52e4b541 req-622a1aeb-3c5e-4dbc-9b6a-506ea1cdada2 service nova] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Received unexpected event network-vif-plugged-57d52bb3-5c0a-4ea0-89b3-94d652923656 for instance with vm_state building and task_state spawning. [ 1049.946947] env[62569]: DEBUG oslo_concurrency.lockutils [None req-ea5e93c3-7ab5-4bc0-b8e2-b1323354386a tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "6e3b1490-830d-4869-b805-b83ef0cf9cc7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.474s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.230058] env[62569]: DEBUG oslo_vmware.api [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250673, 'name': ReconfigVM_Task, 'duration_secs': 0.151023} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.230058] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Reconfigured VM instance instance-0000006a to detach disk 2000 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1050.230646] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-376363b8-18ab-4687-95a2-a2c4d073170e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.257036] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232/f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1050.257336] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dc000523-34a2-40b9-af42-b86ee3973a78 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.274881] env[62569]: DEBUG oslo_vmware.api [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1050.274881] env[62569]: value = "task-1250675" [ 1050.274881] env[62569]: _type = "Task" [ 1050.274881] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.282793] env[62569]: DEBUG oslo_vmware.api [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250675, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.508021] env[62569]: DEBUG nova.network.neutron [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Successfully updated port: 57d52bb3-5c0a-4ea0-89b3-94d652923656 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1050.531833] env[62569]: DEBUG nova.compute.manager [req-a7e8320c-a8b0-44ad-a679-a962eeadcbb2 req-14d70ce3-61ed-46c8-acfb-75b80caf40e6 service nova] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Received event network-changed-57d52bb3-5c0a-4ea0-89b3-94d652923656 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1050.532082] env[62569]: DEBUG nova.compute.manager [req-a7e8320c-a8b0-44ad-a679-a962eeadcbb2 req-14d70ce3-61ed-46c8-acfb-75b80caf40e6 service nova] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Refreshing instance network info cache due to event network-changed-57d52bb3-5c0a-4ea0-89b3-94d652923656. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1050.532329] env[62569]: DEBUG oslo_concurrency.lockutils [req-a7e8320c-a8b0-44ad-a679-a962eeadcbb2 req-14d70ce3-61ed-46c8-acfb-75b80caf40e6 service nova] Acquiring lock "refresh_cache-cf077eb8-8860-46f0-ae84-eed58c29873b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.532527] env[62569]: DEBUG oslo_concurrency.lockutils [req-a7e8320c-a8b0-44ad-a679-a962eeadcbb2 req-14d70ce3-61ed-46c8-acfb-75b80caf40e6 service nova] Acquired lock "refresh_cache-cf077eb8-8860-46f0-ae84-eed58c29873b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.532647] env[62569]: DEBUG nova.network.neutron [req-a7e8320c-a8b0-44ad-a679-a962eeadcbb2 req-14d70ce3-61ed-46c8-acfb-75b80caf40e6 service nova] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Refreshing network info cache for port 57d52bb3-5c0a-4ea0-89b3-94d652923656 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1050.575839] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "ba73d02c-4c4f-441f-8142-f00af4524758" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.576127] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "ba73d02c-4c4f-441f-8142-f00af4524758" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.784410] env[62569]: DEBUG oslo_vmware.api [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250675, 'name': ReconfigVM_Task, 'duration_secs': 0.236762} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.784718] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Reconfigured VM instance instance-0000006a to attach disk [datastore1] f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232/f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1050.785029] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Updating instance 'f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232' progress to 50 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1051.013058] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "refresh_cache-cf077eb8-8860-46f0-ae84-eed58c29873b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.075229] env[62569]: DEBUG nova.network.neutron [req-a7e8320c-a8b0-44ad-a679-a962eeadcbb2 req-14d70ce3-61ed-46c8-acfb-75b80caf40e6 service nova] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1051.078720] env[62569]: DEBUG nova.compute.manager [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1051.153945] env[62569]: DEBUG nova.network.neutron [req-a7e8320c-a8b0-44ad-a679-a962eeadcbb2 req-14d70ce3-61ed-46c8-acfb-75b80caf40e6 service nova] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.291387] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30053117-4730-4224-a3f7-7809a72f2d7b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.310395] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86b8f298-56ea-4ba0-8b5d-a3b352ca5de3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.327951] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Updating instance 'f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232' progress to 67 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1051.599884] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.600180] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.275798] env[62569]: INFO nova.compute.claims [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1052.279313] env[62569]: DEBUG oslo_concurrency.lockutils [req-a7e8320c-a8b0-44ad-a679-a962eeadcbb2 req-14d70ce3-61ed-46c8-acfb-75b80caf40e6 service nova] Releasing lock "refresh_cache-cf077eb8-8860-46f0-ae84-eed58c29873b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.284585] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquired lock "refresh_cache-cf077eb8-8860-46f0-ae84-eed58c29873b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.284828] env[62569]: DEBUG nova.network.neutron [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1052.324284] env[62569]: DEBUG nova.network.neutron [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Port 04dded5b-c41d-4dd5-b9bb-354115b85a5d binding to destination host cpu-1 is already ACTIVE {{(pid=62569) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1052.397893] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-c244e362-a91f-4ea1-882a-472f1b2f7161 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Volume attach. Driver type: vmdk {{(pid=62569) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1052.398164] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-c244e362-a91f-4ea1-882a-472f1b2f7161 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269507', 'volume_id': 'aa9283c3-c013-498f-8986-2367724e045f', 'name': 'volume-aa9283c3-c013-498f-8986-2367724e045f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '659cc0b8-0f8e-4aa3-83ca-d8c04f355d81', 'attached_at': '', 'detached_at': '', 'volume_id': 'aa9283c3-c013-498f-8986-2367724e045f', 'serial': 'aa9283c3-c013-498f-8986-2367724e045f'} {{(pid=62569) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1052.399053] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f7286f-e10d-4915-bc6b-5cc29688f8d3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.414906] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab9e927-53eb-45aa-a174-6d070e0b27dd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.438759] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-c244e362-a91f-4ea1-882a-472f1b2f7161 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] volume-aa9283c3-c013-498f-8986-2367724e045f/volume-aa9283c3-c013-498f-8986-2367724e045f.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1052.438992] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-51e73bee-928d-4511-b13d-97f039612f4a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.456291] env[62569]: DEBUG oslo_vmware.api [None req-c244e362-a91f-4ea1-882a-472f1b2f7161 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1052.456291] env[62569]: value = "task-1250676" [ 1052.456291] env[62569]: _type = "Task" [ 1052.456291] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.463575] env[62569]: DEBUG oslo_vmware.api [None req-c244e362-a91f-4ea1-882a-472f1b2f7161 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250676, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.816913] env[62569]: DEBUG nova.network.neutron [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1052.938414] env[62569]: DEBUG nova.network.neutron [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Updating instance_info_cache with network_info: [{"id": "57d52bb3-5c0a-4ea0-89b3-94d652923656", "address": "fa:16:3e:22:22:04", "network": {"id": "334a2e22-0249-48e3-a2f5-ac37df9771d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1048653043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f7fb3e698324f8e851aefbee74ac669", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f880ac2e-d532-4f54-87bb-998a8d1bca78", "external-id": "nsx-vlan-transportzone-491", "segmentation_id": 491, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57d52bb3-5c", "ovs_interfaceid": "57d52bb3-5c0a-4ea0-89b3-94d652923656", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.965577] env[62569]: DEBUG oslo_vmware.api [None req-c244e362-a91f-4ea1-882a-472f1b2f7161 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250676, 'name': ReconfigVM_Task, 'duration_secs': 0.377596} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.965862] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-c244e362-a91f-4ea1-882a-472f1b2f7161 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Reconfigured VM instance instance-00000066 to attach disk [datastore1] volume-aa9283c3-c013-498f-8986-2367724e045f/volume-aa9283c3-c013-498f-8986-2367724e045f.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1052.970947] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-02dd787c-bd38-4c15-a120-3f36911e93d6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.985499] env[62569]: DEBUG oslo_vmware.api [None req-c244e362-a91f-4ea1-882a-472f1b2f7161 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1052.985499] env[62569]: value = "task-1250677" [ 1052.985499] env[62569]: _type = "Task" [ 1052.985499] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.993244] env[62569]: DEBUG oslo_vmware.api [None req-c244e362-a91f-4ea1-882a-472f1b2f7161 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250677, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.348367] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.348633] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.348768] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.395012] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4de488f-3701-4494-9449-b03ebd6e2ae4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.404341] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6b57289-9f0f-46f7-be8b-195b463e0319 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.433813] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a762f4f7-4531-4afd-8169-173e30233886 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.440318] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Releasing lock "refresh_cache-cf077eb8-8860-46f0-ae84-eed58c29873b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.440604] env[62569]: DEBUG nova.compute.manager [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Instance network_info: |[{"id": "57d52bb3-5c0a-4ea0-89b3-94d652923656", "address": "fa:16:3e:22:22:04", "network": {"id": "334a2e22-0249-48e3-a2f5-ac37df9771d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1048653043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f7fb3e698324f8e851aefbee74ac669", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f880ac2e-d532-4f54-87bb-998a8d1bca78", "external-id": "nsx-vlan-transportzone-491", "segmentation_id": 491, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57d52bb3-5c", "ovs_interfaceid": "57d52bb3-5c0a-4ea0-89b3-94d652923656", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1053.441773] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eb88f34-b831-4383-8f63-951ede32fcd5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.445548] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:22:04', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f880ac2e-d532-4f54-87bb-998a8d1bca78', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '57d52bb3-5c0a-4ea0-89b3-94d652923656', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1053.452717] env[62569]: DEBUG oslo.service.loopingcall [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1053.452920] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1053.453465] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6f8b9c72-3e3e-4050-ad81-7cf3f81c8ff8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.474412] env[62569]: DEBUG nova.compute.provider_tree [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.480722] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1053.480722] env[62569]: value = "task-1250678" [ 1053.480722] env[62569]: _type = "Task" [ 1053.480722] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.488384] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250678, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.495219] env[62569]: DEBUG oslo_vmware.api [None req-c244e362-a91f-4ea1-882a-472f1b2f7161 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250677, 'name': ReconfigVM_Task, 'duration_secs': 0.145802} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.495497] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-c244e362-a91f-4ea1-882a-472f1b2f7161 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269507', 'volume_id': 'aa9283c3-c013-498f-8986-2367724e045f', 'name': 'volume-aa9283c3-c013-498f-8986-2367724e045f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '659cc0b8-0f8e-4aa3-83ca-d8c04f355d81', 'attached_at': '', 'detached_at': '', 'volume_id': 'aa9283c3-c013-498f-8986-2367724e045f', 'serial': 'aa9283c3-c013-498f-8986-2367724e045f'} {{(pid=62569) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1053.977512] env[62569]: DEBUG nova.scheduler.client.report [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1053.991368] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250678, 'name': CreateVM_Task, 'duration_secs': 0.353677} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.991528] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1053.992188] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.992385] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.992711] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1053.993193] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-affb0702-da03-481e-8cff-877f7d118541 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.000172] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 1054.000172] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a572c2-f87a-cd3b-df66-5eface16ccee" [ 1054.000172] env[62569]: _type = "Task" [ 1054.000172] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.009435] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a572c2-f87a-cd3b-df66-5eface16ccee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.067784] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1054.068068] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1054.068222] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Starting heal instance info cache {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 1054.380783] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "refresh_cache-f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.381031] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired lock "refresh_cache-f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.381175] env[62569]: DEBUG nova.network.neutron [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1054.481870] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.882s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.482481] env[62569]: DEBUG nova.compute.manager [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1054.511160] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a572c2-f87a-cd3b-df66-5eface16ccee, 'name': SearchDatastore_Task, 'duration_secs': 0.010678} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.511417] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.511653] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1054.511885] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.512057] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.512287] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1054.512577] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5c4127ce-44b7-4028-9110-e6cda37d7006 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.521041] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1054.521222] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1054.521893] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32274551-7071-46d5-9c92-f13b2b743196 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.527075] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 1054.527075] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a46e21-93a4-f202-3dff-0d0084ad4086" [ 1054.527075] env[62569]: _type = "Task" [ 1054.527075] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.530342] env[62569]: DEBUG nova.objects.instance [None req-c244e362-a91f-4ea1-882a-472f1b2f7161 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lazy-loading 'flavor' on Instance uuid 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1054.535823] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a46e21-93a4-f202-3dff-0d0084ad4086, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.570138] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Didn't find any instances for network info cache update. {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10394}} [ 1054.570384] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1054.570544] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1054.570690] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1054.570842] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1054.570994] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1054.571165] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1054.571296] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62569) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 1054.571436] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1054.987145] env[62569]: DEBUG nova.compute.utils [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1054.988577] env[62569]: DEBUG nova.compute.manager [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1054.988755] env[62569]: DEBUG nova.network.neutron [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1055.035045] env[62569]: DEBUG oslo_concurrency.lockutils [None req-c244e362-a91f-4ea1-882a-472f1b2f7161 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.235s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.040453] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52a46e21-93a4-f202-3dff-0d0084ad4086, 'name': SearchDatastore_Task, 'duration_secs': 0.007999} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.041791] env[62569]: DEBUG nova.policy [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6dd052494f244fd78050c5d26f497720', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2c6a4814c234d53bbd4844cd482fbae', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 1055.043763] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8f6d482-b9ea-4b9d-97ec-2726e9f82df5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.050309] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 1055.050309] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52696198-9b23-3aa6-7e23-1fe63dafe16a" [ 1055.050309] env[62569]: _type = "Task" [ 1055.050309] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.062327] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52696198-9b23-3aa6-7e23-1fe63dafe16a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.073843] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.073933] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.074306] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.074578] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62569) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1055.075373] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8d8c99f-4a4a-463f-befa-5894478ac50f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.083324] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ed3ee3b-bd42-4630-acdf-863d53956980 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.099413] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304c9c83-9e31-4424-9771-967417bfd845 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.105883] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b93192f-eb5c-4d3c-95c5-994a714acfc6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.136491] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180679MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=62569) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1055.136648] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.136853] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.139954] env[62569]: DEBUG nova.network.neutron [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Updating instance_info_cache with network_info: [{"id": "04dded5b-c41d-4dd5-b9bb-354115b85a5d", "address": "fa:16:3e:50:e2:7f", "network": {"id": "30d3bf4a-8f55-492f-a049-181e29731818", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1244949475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bdba6022e3d4697a336ca28ca4eccec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04dded5b-c4", "ovs_interfaceid": "04dded5b-c41d-4dd5-b9bb-354115b85a5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.492581] env[62569]: DEBUG nova.compute.manager [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1055.517186] env[62569]: DEBUG nova.network.neutron [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Successfully created port: e47e027b-ae9f-4e3d-a913-21e9088d31f0 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1055.564555] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52696198-9b23-3aa6-7e23-1fe63dafe16a, 'name': SearchDatastore_Task, 'duration_secs': 0.009098} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.564871] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.565149] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] cf077eb8-8860-46f0-ae84-eed58c29873b/cf077eb8-8860-46f0-ae84-eed58c29873b.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1055.565403] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c3b007b2-52eb-460e-99a2-a769cd630363 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.572110] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 1055.572110] env[62569]: value = "task-1250679" [ 1055.572110] env[62569]: _type = "Task" [ 1055.572110] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.580013] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250679, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.644022] env[62569]: DEBUG oslo_concurrency.lockutils [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Releasing lock "refresh_cache-f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.932035] env[62569]: DEBUG nova.compute.manager [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Stashing vm_state: active {{(pid=62569) _prep_resize /opt/stack/nova/nova/compute/manager.py:5997}} [ 1056.082084] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250679, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.438462} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.082368] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] cf077eb8-8860-46f0-ae84-eed58c29873b/cf077eb8-8860-46f0-ae84-eed58c29873b.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1056.082615] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1056.082875] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-906f9dcd-c2a7-48c4-aee9-f3b096bf82cc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.089691] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 1056.089691] env[62569]: value = "task-1250680" [ 1056.089691] env[62569]: _type = "Task" [ 1056.089691] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.096993] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250680, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.155997] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Applying migration context for instance f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232 as it has an incoming, in-progress migration de882a97-fdf6-4238-9efb-f038b01e1cf6. Migration status is post-migrating {{(pid=62569) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1056.157800] env[62569]: INFO nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Updating resource usage from migration 56485982-7035-45de-a101-71589195430f [ 1056.158107] env[62569]: INFO nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Updating resource usage from migration de882a97-fdf6-4238-9efb-f038b01e1cf6 [ 1056.179912] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 534c9ece-c8df-4528-83d8-c6c0ef8c4793 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1056.180093] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 96167629-8777-4a1c-b564-c70d1bb59b90 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1056.180224] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Migration de882a97-fdf6-4238-9efb-f038b01e1cf6 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1056.180343] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1056.180500] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance cf077eb8-8860-46f0-ae84-eed58c29873b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1056.180641] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance ba73d02c-4c4f-441f-8142-f00af4524758 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1056.180761] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Migration 56485982-7035-45de-a101-71589195430f is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1056.180980] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1056.181206] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=62569) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1056.181347] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1920MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=62569) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1056.185532] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83384e8d-4882-4fb1-b1fc-3fd738e93066 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.207308] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcb85f4d-dcb6-4caf-9f93-cb817c516a35 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.214453] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Updating instance 'f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232' progress to 83 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1056.306105] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03c19d25-68d0-42a8-990e-e37fba4316b7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.313881] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6547aa50-af62-4af9-8fa2-6f9079fd9ff3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.345851] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-244bc974-2426-453c-8aad-7560b9d12dda {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.353200] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03363042-bbfe-4af7-9041-576bc17281f3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.366471] env[62569]: DEBUG nova.compute.provider_tree [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1056.448825] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.501534] env[62569]: DEBUG nova.compute.manager [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1056.526064] env[62569]: DEBUG nova.virt.hardware [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1056.526336] env[62569]: DEBUG nova.virt.hardware [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1056.526501] env[62569]: DEBUG nova.virt.hardware [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1056.526688] env[62569]: DEBUG nova.virt.hardware [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1056.526839] env[62569]: DEBUG nova.virt.hardware [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1056.526991] env[62569]: DEBUG nova.virt.hardware [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1056.527220] env[62569]: DEBUG nova.virt.hardware [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1056.527385] env[62569]: DEBUG nova.virt.hardware [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1056.527559] env[62569]: DEBUG nova.virt.hardware [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1056.527729] env[62569]: DEBUG nova.virt.hardware [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1056.527906] env[62569]: DEBUG nova.virt.hardware [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1056.528788] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a0af104-d161-4c6e-ada3-9f6120e2c87a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.536976] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba023003-d783-452c-b56f-855c3ad1c62f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.598077] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250680, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.230238} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.598303] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1056.599035] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b55c92d-0197-4b4d-bdcd-0134f00a9c9c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.620358] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] cf077eb8-8860-46f0-ae84-eed58c29873b/cf077eb8-8860-46f0-ae84-eed58c29873b.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1056.620606] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f7cfd80e-af4a-4fa7-a9da-f6cf06d28953 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.639800] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 1056.639800] env[62569]: value = "task-1250681" [ 1056.639800] env[62569]: _type = "Task" [ 1056.639800] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.647075] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250681, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.721384] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1056.721736] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6e17a156-1ba7-420f-960a-0a7bbfa4f295 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.729062] env[62569]: DEBUG oslo_vmware.api [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1056.729062] env[62569]: value = "task-1250682" [ 1056.729062] env[62569]: _type = "Task" [ 1056.729062] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.740272] env[62569]: DEBUG oslo_vmware.api [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250682, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.872017] env[62569]: DEBUG nova.scheduler.client.report [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1057.016166] env[62569]: DEBUG nova.compute.manager [req-db148798-f7d1-4810-9134-eb9707600630 req-ac295075-d1fe-490d-a12c-f6461658ab37 service nova] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Received event network-vif-plugged-e47e027b-ae9f-4e3d-a913-21e9088d31f0 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1057.016166] env[62569]: DEBUG oslo_concurrency.lockutils [req-db148798-f7d1-4810-9134-eb9707600630 req-ac295075-d1fe-490d-a12c-f6461658ab37 service nova] Acquiring lock "ba73d02c-4c4f-441f-8142-f00af4524758-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.016166] env[62569]: DEBUG oslo_concurrency.lockutils [req-db148798-f7d1-4810-9134-eb9707600630 req-ac295075-d1fe-490d-a12c-f6461658ab37 service nova] Lock "ba73d02c-4c4f-441f-8142-f00af4524758-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.016855] env[62569]: DEBUG oslo_concurrency.lockutils [req-db148798-f7d1-4810-9134-eb9707600630 req-ac295075-d1fe-490d-a12c-f6461658ab37 service nova] Lock "ba73d02c-4c4f-441f-8142-f00af4524758-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.017194] env[62569]: DEBUG nova.compute.manager [req-db148798-f7d1-4810-9134-eb9707600630 req-ac295075-d1fe-490d-a12c-f6461658ab37 service nova] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] No waiting events found dispatching network-vif-plugged-e47e027b-ae9f-4e3d-a913-21e9088d31f0 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1057.017479] env[62569]: WARNING nova.compute.manager [req-db148798-f7d1-4810-9134-eb9707600630 req-ac295075-d1fe-490d-a12c-f6461658ab37 service nova] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Received unexpected event network-vif-plugged-e47e027b-ae9f-4e3d-a913-21e9088d31f0 for instance with vm_state building and task_state spawning. [ 1057.152911] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250681, 'name': ReconfigVM_Task, 'duration_secs': 0.368085} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.152911] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Reconfigured VM instance instance-0000006d to attach disk [datastore1] cf077eb8-8860-46f0-ae84-eed58c29873b/cf077eb8-8860-46f0-ae84-eed58c29873b.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1057.152911] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dcdd781e-fe61-41dc-96ab-bb80852961ca {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.159029] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 1057.159029] env[62569]: value = "task-1250683" [ 1057.159029] env[62569]: _type = "Task" [ 1057.159029] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.165387] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250683, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.212518] env[62569]: DEBUG nova.network.neutron [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Successfully updated port: e47e027b-ae9f-4e3d-a913-21e9088d31f0 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1057.239062] env[62569]: DEBUG oslo_vmware.api [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250682, 'name': PowerOnVM_Task, 'duration_secs': 0.392109} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.239389] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1057.239581] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2e6609-48b6-426d-a782-370ee7ab74a9 tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Updating instance 'f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232' progress to 100 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1057.376154] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62569) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1057.376463] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.239s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.376865] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.928s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.655949] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1057.656322] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1057.656322] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Cleaning up deleted instances with incomplete migration {{(pid=62569) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11646}} [ 1057.665866] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250683, 'name': Rename_Task, 'duration_secs': 0.39522} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.666675] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1057.666923] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2d5b7199-5a8a-4d5a-9183-d6c130557ef0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.673547] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 1057.673547] env[62569]: value = "task-1250684" [ 1057.673547] env[62569]: _type = "Task" [ 1057.673547] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.680961] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250684, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.717445] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "refresh_cache-ba73d02c-4c4f-441f-8142-f00af4524758" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.717597] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired lock "refresh_cache-ba73d02c-4c4f-441f-8142-f00af4524758" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.717750] env[62569]: DEBUG nova.network.neutron [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1057.882514] env[62569]: INFO nova.compute.claims [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1058.183854] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250684, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.252285] env[62569]: DEBUG nova.network.neutron [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1058.390832] env[62569]: INFO nova.compute.resource_tracker [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Updating resource usage from migration 56485982-7035-45de-a101-71589195430f [ 1058.474896] env[62569]: DEBUG nova.network.neutron [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Updating instance_info_cache with network_info: [{"id": "e47e027b-ae9f-4e3d-a913-21e9088d31f0", "address": "fa:16:3e:7b:46:46", "network": {"id": "884d3248-07f1-4cf7-8ceb-b94b34622384", "bridge": "br-int", "label": "tempest-ServersTestJSON-1882023778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2c6a4814c234d53bbd4844cd482fbae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape47e027b-ae", "ovs_interfaceid": "e47e027b-ae9f-4e3d-a913-21e9088d31f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.497407] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cf03f69-2e68-48f4-b122-8b03553b35b7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.504883] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-719a4a15-37a4-4009-b12c-547bd54ad8f2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.534418] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b3e6b83-7e51-463b-874a-5ba0bbe4eb81 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.541183] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6ccc646-de75-449c-b3f4-95c62d4298ac {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.553715] env[62569]: DEBUG nova.compute.provider_tree [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1058.684082] env[62569]: DEBUG oslo_vmware.api [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250684, 'name': PowerOnVM_Task, 'duration_secs': 0.510637} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.684455] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1058.684691] env[62569]: INFO nova.compute.manager [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Took 8.92 seconds to spawn the instance on the hypervisor. [ 1058.684901] env[62569]: DEBUG nova.compute.manager [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1058.685659] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e587803-5502-40a9-8bcb-8da4b05f9bd6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.979899] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Releasing lock "refresh_cache-ba73d02c-4c4f-441f-8142-f00af4524758" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.980274] env[62569]: DEBUG nova.compute.manager [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Instance network_info: |[{"id": "e47e027b-ae9f-4e3d-a913-21e9088d31f0", "address": "fa:16:3e:7b:46:46", "network": {"id": "884d3248-07f1-4cf7-8ceb-b94b34622384", "bridge": "br-int", "label": "tempest-ServersTestJSON-1882023778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2c6a4814c234d53bbd4844cd482fbae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape47e027b-ae", "ovs_interfaceid": "e47e027b-ae9f-4e3d-a913-21e9088d31f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1058.980762] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7b:46:46', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e9fa4744-8702-4973-b911-ee18192a3e4b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e47e027b-ae9f-4e3d-a913-21e9088d31f0', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1058.988749] env[62569]: DEBUG oslo.service.loopingcall [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1058.988990] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1058.989399] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-019957c1-ed15-4b13-8b12-70b3de6679b4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.009060] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1059.009060] env[62569]: value = "task-1250685" [ 1059.009060] env[62569]: _type = "Task" [ 1059.009060] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.018432] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250685, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.048067] env[62569]: DEBUG nova.compute.manager [req-9a4ace58-f2ad-4795-aa7d-f7af37f85d74 req-ef1c7863-5725-4cd5-ab04-1d6cdf216a35 service nova] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Received event network-changed-e47e027b-ae9f-4e3d-a913-21e9088d31f0 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1059.048281] env[62569]: DEBUG nova.compute.manager [req-9a4ace58-f2ad-4795-aa7d-f7af37f85d74 req-ef1c7863-5725-4cd5-ab04-1d6cdf216a35 service nova] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Refreshing instance network info cache due to event network-changed-e47e027b-ae9f-4e3d-a913-21e9088d31f0. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1059.048506] env[62569]: DEBUG oslo_concurrency.lockutils [req-9a4ace58-f2ad-4795-aa7d-f7af37f85d74 req-ef1c7863-5725-4cd5-ab04-1d6cdf216a35 service nova] Acquiring lock "refresh_cache-ba73d02c-4c4f-441f-8142-f00af4524758" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.048654] env[62569]: DEBUG oslo_concurrency.lockutils [req-9a4ace58-f2ad-4795-aa7d-f7af37f85d74 req-ef1c7863-5725-4cd5-ab04-1d6cdf216a35 service nova] Acquired lock "refresh_cache-ba73d02c-4c4f-441f-8142-f00af4524758" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.048818] env[62569]: DEBUG nova.network.neutron [req-9a4ace58-f2ad-4795-aa7d-f7af37f85d74 req-ef1c7863-5725-4cd5-ab04-1d6cdf216a35 service nova] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Refreshing network info cache for port e47e027b-ae9f-4e3d-a913-21e9088d31f0 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1059.056245] env[62569]: DEBUG nova.scheduler.client.report [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1059.161500] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1059.205346] env[62569]: INFO nova.compute.manager [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Took 14.33 seconds to build instance. [ 1059.520265] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250685, 'name': CreateVM_Task, 'duration_secs': 0.299807} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.520483] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1059.521259] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.521458] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.521825] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1059.522131] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4f3b8c7-1e82-4caf-a260-47d99af3b52c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.526834] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1059.526834] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]524e90b1-15bc-297f-41f9-281c19468298" [ 1059.526834] env[62569]: _type = "Task" [ 1059.526834] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.534918] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]524e90b1-15bc-297f-41f9-281c19468298, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.561490] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.185s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.561703] env[62569]: INFO nova.compute.manager [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Migrating [ 1059.656232] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1059.656411] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Starting heal instance info cache {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 1059.656528] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Rebuilding the list of instances to heal {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10312}} [ 1059.707735] env[62569]: DEBUG oslo_concurrency.lockutils [None req-8638dc2e-4af9-4857-80ee-eccce4bf9c1d tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cf077eb8-8860-46f0-ae84-eed58c29873b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.847s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.868656] env[62569]: DEBUG nova.network.neutron [req-9a4ace58-f2ad-4795-aa7d-f7af37f85d74 req-ef1c7863-5725-4cd5-ab04-1d6cdf216a35 service nova] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Updated VIF entry in instance network info cache for port e47e027b-ae9f-4e3d-a913-21e9088d31f0. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1059.869054] env[62569]: DEBUG nova.network.neutron [req-9a4ace58-f2ad-4795-aa7d-f7af37f85d74 req-ef1c7863-5725-4cd5-ab04-1d6cdf216a35 service nova] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Updating instance_info_cache with network_info: [{"id": "e47e027b-ae9f-4e3d-a913-21e9088d31f0", "address": "fa:16:3e:7b:46:46", "network": {"id": "884d3248-07f1-4cf7-8ceb-b94b34622384", "bridge": "br-int", "label": "tempest-ServersTestJSON-1882023778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2c6a4814c234d53bbd4844cd482fbae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape47e027b-ae", "ovs_interfaceid": "e47e027b-ae9f-4e3d-a913-21e9088d31f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.039318] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]524e90b1-15bc-297f-41f9-281c19468298, 'name': SearchDatastore_Task, 'duration_secs': 0.009999} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.039651] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.039900] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1060.040162] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.040317] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.040524] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1060.040790] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c3b1a50e-2203-4b7f-b03a-5c0f0d00bb8b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.048781] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1060.048962] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1060.049662] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c19bfca-5e44-4d6c-a921-6b2d7f722b6f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.056139] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1060.056139] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]526bb58d-e176-2a66-eb4d-3da97dd9d26d" [ 1060.056139] env[62569]: _type = "Task" [ 1060.056139] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.063469] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]526bb58d-e176-2a66-eb4d-3da97dd9d26d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.078349] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "refresh_cache-659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.078544] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired lock "refresh_cache-659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.078696] env[62569]: DEBUG nova.network.neutron [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1060.161080] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Skipping network cache update for instance because it is Building. {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10321}} [ 1060.188923] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "refresh_cache-534c9ece-c8df-4528-83d8-c6c0ef8c4793" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.189091] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquired lock "refresh_cache-534c9ece-c8df-4528-83d8-c6c0ef8c4793" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.189242] env[62569]: DEBUG nova.network.neutron [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Forcefully refreshing network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1060.189394] env[62569]: DEBUG nova.objects.instance [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lazy-loading 'info_cache' on Instance uuid 534c9ece-c8df-4528-83d8-c6c0ef8c4793 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1060.229161] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.229414] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.229597] env[62569]: DEBUG nova.compute.manager [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Going to confirm migration 5 {{(pid=62569) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5156}} [ 1060.371635] env[62569]: DEBUG oslo_concurrency.lockutils [req-9a4ace58-f2ad-4795-aa7d-f7af37f85d74 req-ef1c7863-5725-4cd5-ab04-1d6cdf216a35 service nova] Releasing lock "refresh_cache-ba73d02c-4c4f-441f-8142-f00af4524758" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.566210] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]526bb58d-e176-2a66-eb4d-3da97dd9d26d, 'name': SearchDatastore_Task, 'duration_secs': 0.008021} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.566975] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2742d70-c91f-4048-8ac0-630e622859c5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.571796] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1060.571796] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52284913-7295-4af5-7400-8439f43f352a" [ 1060.571796] env[62569]: _type = "Task" [ 1060.571796] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.578832] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52284913-7295-4af5-7400-8439f43f352a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.799031] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "refresh_cache-f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.799272] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquired lock "refresh_cache-f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.799428] env[62569]: DEBUG nova.network.neutron [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1060.799622] env[62569]: DEBUG nova.objects.instance [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lazy-loading 'info_cache' on Instance uuid f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1060.846152] env[62569]: DEBUG nova.network.neutron [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Updating instance_info_cache with network_info: [{"id": "8b78baca-9288-4ab0-b21d-7fb96baeaf42", "address": "fa:16:3e:7f:fc:2d", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b78baca-92", "ovs_interfaceid": "8b78baca-9288-4ab0-b21d-7fb96baeaf42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.073751] env[62569]: DEBUG nova.compute.manager [req-3d05947a-2c80-479b-a297-e48ffc394ea8 req-10f18388-d8fd-4906-87b7-e1efee6ecd62 service nova] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Received event network-changed-57d52bb3-5c0a-4ea0-89b3-94d652923656 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1061.073965] env[62569]: DEBUG nova.compute.manager [req-3d05947a-2c80-479b-a297-e48ffc394ea8 req-10f18388-d8fd-4906-87b7-e1efee6ecd62 service nova] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Refreshing instance network info cache due to event network-changed-57d52bb3-5c0a-4ea0-89b3-94d652923656. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1061.074198] env[62569]: DEBUG oslo_concurrency.lockutils [req-3d05947a-2c80-479b-a297-e48ffc394ea8 req-10f18388-d8fd-4906-87b7-e1efee6ecd62 service nova] Acquiring lock "refresh_cache-cf077eb8-8860-46f0-ae84-eed58c29873b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.074346] env[62569]: DEBUG oslo_concurrency.lockutils [req-3d05947a-2c80-479b-a297-e48ffc394ea8 req-10f18388-d8fd-4906-87b7-e1efee6ecd62 service nova] Acquired lock "refresh_cache-cf077eb8-8860-46f0-ae84-eed58c29873b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.074508] env[62569]: DEBUG nova.network.neutron [req-3d05947a-2c80-479b-a297-e48ffc394ea8 req-10f18388-d8fd-4906-87b7-e1efee6ecd62 service nova] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Refreshing network info cache for port 57d52bb3-5c0a-4ea0-89b3-94d652923656 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1061.085138] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52284913-7295-4af5-7400-8439f43f352a, 'name': SearchDatastore_Task, 'duration_secs': 0.010616} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.085819] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.086083] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] ba73d02c-4c4f-441f-8142-f00af4524758/ba73d02c-4c4f-441f-8142-f00af4524758.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1061.086335] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-961dfef6-994d-44bc-b294-5c7930a9c1d8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.093271] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1061.093271] env[62569]: value = "task-1250686" [ 1061.093271] env[62569]: _type = "Task" [ 1061.093271] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.100893] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250686, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.348641] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Releasing lock "refresh_cache-659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.603019] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250686, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.441386} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.603232] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] ba73d02c-4c4f-441f-8142-f00af4524758/ba73d02c-4c4f-441f-8142-f00af4524758.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1061.603448] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1061.603795] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e6f756bc-91f4-41f9-b06a-6bd51c54f164 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.611180] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1061.611180] env[62569]: value = "task-1250687" [ 1061.611180] env[62569]: _type = "Task" [ 1061.611180] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.618818] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250687, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.785607] env[62569]: DEBUG nova.network.neutron [req-3d05947a-2c80-479b-a297-e48ffc394ea8 req-10f18388-d8fd-4906-87b7-e1efee6ecd62 service nova] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Updated VIF entry in instance network info cache for port 57d52bb3-5c0a-4ea0-89b3-94d652923656. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1061.785995] env[62569]: DEBUG nova.network.neutron [req-3d05947a-2c80-479b-a297-e48ffc394ea8 req-10f18388-d8fd-4906-87b7-e1efee6ecd62 service nova] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Updating instance_info_cache with network_info: [{"id": "57d52bb3-5c0a-4ea0-89b3-94d652923656", "address": "fa:16:3e:22:22:04", "network": {"id": "334a2e22-0249-48e3-a2f5-ac37df9771d4", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1048653043-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f7fb3e698324f8e851aefbee74ac669", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f880ac2e-d532-4f54-87bb-998a8d1bca78", "external-id": "nsx-vlan-transportzone-491", "segmentation_id": 491, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57d52bb3-5c", "ovs_interfaceid": "57d52bb3-5c0a-4ea0-89b3-94d652923656", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.916961] env[62569]: DEBUG nova.network.neutron [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Updating instance_info_cache with network_info: [{"id": "37c7a902-07de-45dc-b683-311b84a5edc5", "address": "fa:16:3e:e3:66:93", "network": {"id": "884d3248-07f1-4cf7-8ceb-b94b34622384", "bridge": "br-int", "label": "tempest-ServersTestJSON-1882023778-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2c6a4814c234d53bbd4844cd482fbae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e9fa4744-8702-4973-b911-ee18192a3e4b", "external-id": "nsx-vlan-transportzone-318", "segmentation_id": 318, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37c7a902-07", "ovs_interfaceid": "37c7a902-07de-45dc-b683-311b84a5edc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.041671] env[62569]: DEBUG nova.network.neutron [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Updating instance_info_cache with network_info: [{"id": "04dded5b-c41d-4dd5-b9bb-354115b85a5d", "address": "fa:16:3e:50:e2:7f", "network": {"id": "30d3bf4a-8f55-492f-a049-181e29731818", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1244949475-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bdba6022e3d4697a336ca28ca4eccec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "983826cf-6390-4ec6-bf97-30a1060947fc", "external-id": "nsx-vlan-transportzone-367", "segmentation_id": 367, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04dded5b-c4", "ovs_interfaceid": "04dded5b-c41d-4dd5-b9bb-354115b85a5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.121529] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250687, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.305325} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.121860] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1062.122676] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a73e106c-fb27-4007-8e10-070dcf7cb839 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.143945] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] ba73d02c-4c4f-441f-8142-f00af4524758/ba73d02c-4c4f-441f-8142-f00af4524758.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1062.144229] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-791eb8b1-2cab-4be0-aa94-0a6104c622ce {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.162772] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1062.162772] env[62569]: value = "task-1250688" [ 1062.162772] env[62569]: _type = "Task" [ 1062.162772] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.170041] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250688, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.290933] env[62569]: DEBUG oslo_concurrency.lockutils [req-3d05947a-2c80-479b-a297-e48ffc394ea8 req-10f18388-d8fd-4906-87b7-e1efee6ecd62 service nova] Releasing lock "refresh_cache-cf077eb8-8860-46f0-ae84-eed58c29873b" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.423256] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Releasing lock "refresh_cache-534c9ece-c8df-4528-83d8-c6c0ef8c4793" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.423439] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Updated the network info_cache for instance {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10379}} [ 1062.427362] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1062.427362] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1062.427362] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1062.427362] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1062.427362] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62569) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 1062.427362] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1062.544529] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Releasing lock "refresh_cache-f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.544815] env[62569]: DEBUG nova.objects.instance [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lazy-loading 'migration_context' on Instance uuid f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1062.672382] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250688, 'name': ReconfigVM_Task, 'duration_secs': 0.364199} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.672639] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Reconfigured VM instance instance-0000006e to attach disk [datastore1] ba73d02c-4c4f-441f-8142-f00af4524758/ba73d02c-4c4f-441f-8142-f00af4524758.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1062.673250] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4085b2d7-e957-4f2f-bc95-b983d8944228 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.679273] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1062.679273] env[62569]: value = "task-1250689" [ 1062.679273] env[62569]: _type = "Task" [ 1062.679273] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.687594] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250689, 'name': Rename_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.868092] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04536005-def7-4601-99de-6cd11b9c3eca {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.888512] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Updating instance '659cc0b8-0f8e-4aa3-83ca-d8c04f355d81' progress to 0 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1062.930164] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.930453] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.930544] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.930693] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62569) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1062.931797] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a27cce-7c41-4da5-a979-c008f26a3850 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.940250] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6c8211f-b767-46ab-b328-6b87dc9c1331 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.953712] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2150de05-fa22-4fc2-833c-0d05d341eacf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.959603] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73a4769e-24a9-4864-875f-574249c8fed5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.987861] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180679MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=62569) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1062.988012] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.988224] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.048093] env[62569]: DEBUG nova.objects.base [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62569) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1063.049080] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8666bcb5-04c1-4228-aa28-a0d9cff1684b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.069380] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4b10040-5f7c-4afd-8e94-2cc8890713ca {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.074645] env[62569]: DEBUG oslo_vmware.api [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1063.074645] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52877110-2eee-ba26-35f3-45d140b676f5" [ 1063.074645] env[62569]: _type = "Task" [ 1063.074645] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.081869] env[62569]: DEBUG oslo_vmware.api [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52877110-2eee-ba26-35f3-45d140b676f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.189506] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250689, 'name': Rename_Task, 'duration_secs': 0.136744} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.189743] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1063.190044] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-29cac160-553a-40d5-b9a8-d2294a27fb4a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.196311] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1063.196311] env[62569]: value = "task-1250690" [ 1063.196311] env[62569]: _type = "Task" [ 1063.196311] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.204304] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250690, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.394455] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1063.394845] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b0e4b1e5-9313-4896-af30-f5f0d2305b69 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.401451] env[62569]: DEBUG oslo_vmware.api [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1063.401451] env[62569]: value = "task-1250691" [ 1063.401451] env[62569]: _type = "Task" [ 1063.401451] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.410910] env[62569]: DEBUG oslo_vmware.api [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250691, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.583975] env[62569]: DEBUG oslo_vmware.api [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52877110-2eee-ba26-35f3-45d140b676f5, 'name': SearchDatastore_Task, 'duration_secs': 0.008889} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.584373] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.706597] env[62569]: DEBUG oslo_vmware.api [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250690, 'name': PowerOnVM_Task, 'duration_secs': 0.494745} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.706877] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1063.707102] env[62569]: INFO nova.compute.manager [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Took 7.21 seconds to spawn the instance on the hypervisor. [ 1063.707297] env[62569]: DEBUG nova.compute.manager [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1063.708073] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93925f1d-f413-4867-8f2d-fd0cc295412d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.911340] env[62569]: DEBUG oslo_vmware.api [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250691, 'name': PowerOffVM_Task, 'duration_secs': 0.261704} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.911627] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1063.911821] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Updating instance '659cc0b8-0f8e-4aa3-83ca-d8c04f355d81' progress to 17 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1063.996755] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Applying migration context for instance 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81 as it has an incoming, in-progress migration 56485982-7035-45de-a101-71589195430f. Migration status is migrating {{(pid=62569) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1063.997102] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Applying migration context for instance f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232 as it has an incoming, in-progress migration de882a97-fdf6-4238-9efb-f038b01e1cf6. Migration status is finished {{(pid=62569) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1063.997992] env[62569]: INFO nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Updating resource usage from migration 56485982-7035-45de-a101-71589195430f [ 1063.998338] env[62569]: INFO nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Updating resource usage from migration de882a97-fdf6-4238-9efb-f038b01e1cf6 [ 1064.093615] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 534c9ece-c8df-4528-83d8-c6c0ef8c4793 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.093739] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 96167629-8777-4a1c-b564-c70d1bb59b90 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.093861] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Migration de882a97-fdf6-4238-9efb-f038b01e1cf6 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1064.093985] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.094361] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance cf077eb8-8860-46f0-ae84-eed58c29873b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.094508] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance ba73d02c-4c4f-441f-8142-f00af4524758 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.094637] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Migration 56485982-7035-45de-a101-71589195430f is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1064.094817] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1064.094972] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=62569) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1064.095135] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2176MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=62569) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1064.194821] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10c7ba88-f76a-4b6d-b9f1-8272968a4f80 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.202698] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8222ad73-4898-4b21-bf9a-065babc54840 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.237370] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c9bd98-6cda-4482-9869-6f97f510d086 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.243765] env[62569]: INFO nova.compute.manager [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Took 12.66 seconds to build instance. [ 1064.247973] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f65f57b2-66f6-4db5-86ad-7e31ef5d4b50 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.261213] env[62569]: DEBUG nova.compute.provider_tree [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.418516] env[62569]: DEBUG nova.virt.hardware [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1064.418793] env[62569]: DEBUG nova.virt.hardware [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1064.419084] env[62569]: DEBUG nova.virt.hardware [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1064.419249] env[62569]: DEBUG nova.virt.hardware [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1064.419404] env[62569]: DEBUG nova.virt.hardware [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1064.420032] env[62569]: DEBUG nova.virt.hardware [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1064.420032] env[62569]: DEBUG nova.virt.hardware [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1064.420032] env[62569]: DEBUG nova.virt.hardware [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1064.420171] env[62569]: DEBUG nova.virt.hardware [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1064.420264] env[62569]: DEBUG nova.virt.hardware [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1064.422931] env[62569]: DEBUG nova.virt.hardware [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1064.425626] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e4e175a-49a2-49df-b839-181276a19db1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.441170] env[62569]: DEBUG oslo_vmware.api [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1064.441170] env[62569]: value = "task-1250692" [ 1064.441170] env[62569]: _type = "Task" [ 1064.441170] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.450487] env[62569]: DEBUG oslo_vmware.api [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250692, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.748325] env[62569]: DEBUG oslo_concurrency.lockutils [None req-9ac04a14-f3b8-4623-bd27-e0f3d00af6c8 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "ba73d02c-4c4f-441f-8142-f00af4524758" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.172s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.764540] env[62569]: DEBUG nova.scheduler.client.report [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1064.953171] env[62569]: DEBUG oslo_vmware.api [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250692, 'name': ReconfigVM_Task, 'duration_secs': 0.170229} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.953761] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Updating instance '659cc0b8-0f8e-4aa3-83ca-d8c04f355d81' progress to 33 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1065.269873] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62569) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1065.270155] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.282s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.270435] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.686s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.271834] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1065.272030] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Cleaning up deleted instances {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11608}} [ 1065.381791] env[62569]: DEBUG oslo_concurrency.lockutils [None req-52fe6e5e-28c9-4ed3-b141-fcbe186a6ece tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "ba73d02c-4c4f-441f-8142-f00af4524758" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.382083] env[62569]: DEBUG oslo_concurrency.lockutils [None req-52fe6e5e-28c9-4ed3-b141-fcbe186a6ece tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "ba73d02c-4c4f-441f-8142-f00af4524758" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.382279] env[62569]: DEBUG nova.compute.manager [None req-52fe6e5e-28c9-4ed3-b141-fcbe186a6ece tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1065.383383] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94d4c5c1-ae0d-46fa-bfe1-07eb023fb608 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.390178] env[62569]: DEBUG nova.compute.manager [None req-52fe6e5e-28c9-4ed3-b141-fcbe186a6ece tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62569) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1065.390733] env[62569]: DEBUG nova.objects.instance [None req-52fe6e5e-28c9-4ed3-b141-fcbe186a6ece tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lazy-loading 'flavor' on Instance uuid ba73d02c-4c4f-441f-8142-f00af4524758 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1065.460102] env[62569]: DEBUG nova.virt.hardware [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1065.460385] env[62569]: DEBUG nova.virt.hardware [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1065.460568] env[62569]: DEBUG nova.virt.hardware [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1065.460789] env[62569]: DEBUG nova.virt.hardware [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1065.460974] env[62569]: DEBUG nova.virt.hardware [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1065.461151] env[62569]: DEBUG nova.virt.hardware [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1065.461361] env[62569]: DEBUG nova.virt.hardware [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1065.461523] env[62569]: DEBUG nova.virt.hardware [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1065.461692] env[62569]: DEBUG nova.virt.hardware [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1065.461860] env[62569]: DEBUG nova.virt.hardware [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1065.462048] env[62569]: DEBUG nova.virt.hardware [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1065.467486] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Reconfiguring VM instance instance-00000066 to detach disk 2000 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1065.467772] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7fd234e8-424d-46ef-b26e-e583760a0c0d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.486269] env[62569]: DEBUG oslo_vmware.api [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1065.486269] env[62569]: value = "task-1250693" [ 1065.486269] env[62569]: _type = "Task" [ 1065.486269] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.494612] env[62569]: DEBUG oslo_vmware.api [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250693, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.789352] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] There are 54 instances to clean {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11617}} [ 1065.789601] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 6e3b1490-830d-4869-b805-b83ef0cf9cc7] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1065.896803] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34868b93-32fd-4481-a55a-3d9094d66229 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.905064] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f3e416-c6db-491e-bdb2-1144cf66cef1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.935230] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab978481-90a8-4819-9385-e5e0e0c08d46 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.942289] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebe1bb3c-616f-49d4-a62a-7a78da40fe20 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.956321] env[62569]: DEBUG nova.compute.provider_tree [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1065.997058] env[62569]: DEBUG oslo_vmware.api [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250693, 'name': ReconfigVM_Task, 'duration_secs': 0.246563} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.997058] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Reconfigured VM instance instance-00000066 to detach disk 2000 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1065.997862] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8547610f-5345-4698-9a99-0f674a9533b4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.023896] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81/659cc0b8-0f8e-4aa3-83ca-d8c04f355d81.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1066.024612] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05237958-eb98-4249-8f19-ce7038b494ee {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.042169] env[62569]: DEBUG oslo_vmware.api [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1066.042169] env[62569]: value = "task-1250694" [ 1066.042169] env[62569]: _type = "Task" [ 1066.042169] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.049577] env[62569]: DEBUG oslo_vmware.api [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250694, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.292996] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 0be076c2-176f-47a9-9489-2f7dfe6d1ce3] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1066.401987] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-52fe6e5e-28c9-4ed3-b141-fcbe186a6ece tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1066.402383] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-28c29e02-8daa-4e40-9f3a-f10ead43c259 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.410446] env[62569]: DEBUG oslo_vmware.api [None req-52fe6e5e-28c9-4ed3-b141-fcbe186a6ece tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1066.410446] env[62569]: value = "task-1250695" [ 1066.410446] env[62569]: _type = "Task" [ 1066.410446] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.418746] env[62569]: DEBUG oslo_vmware.api [None req-52fe6e5e-28c9-4ed3-b141-fcbe186a6ece tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250695, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.459687] env[62569]: DEBUG nova.scheduler.client.report [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1066.551470] env[62569]: DEBUG oslo_vmware.api [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250694, 'name': ReconfigVM_Task, 'duration_secs': 0.274125} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.551710] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81/659cc0b8-0f8e-4aa3-83ca-d8c04f355d81.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1066.552010] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Updating instance '659cc0b8-0f8e-4aa3-83ca-d8c04f355d81' progress to 50 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1066.796378] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: c896f38c-f313-491b-86c5-a785721667d7] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1066.920611] env[62569]: DEBUG oslo_vmware.api [None req-52fe6e5e-28c9-4ed3-b141-fcbe186a6ece tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250695, 'name': PowerOffVM_Task, 'duration_secs': 0.281851} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.920852] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-52fe6e5e-28c9-4ed3-b141-fcbe186a6ece tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1066.921044] env[62569]: DEBUG nova.compute.manager [None req-52fe6e5e-28c9-4ed3-b141-fcbe186a6ece tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1066.921788] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc849bc9-d054-4313-a9b6-4453380ebe27 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.058763] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba15c0c5-d8cd-4e2f-8e21-d6ace0f7cb44 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.080948] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-016deafd-61a7-4eff-82f4-a1e533d90070 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.102269] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Updating instance '659cc0b8-0f8e-4aa3-83ca-d8c04f355d81' progress to 67 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1067.299795] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 4bf2f0fb-420f-4c13-97c1-974af7c49658] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1067.432525] env[62569]: DEBUG oslo_concurrency.lockutils [None req-52fe6e5e-28c9-4ed3-b141-fcbe186a6ece tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "ba73d02c-4c4f-441f-8142-f00af4524758" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.050s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.469635] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.199s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.803060] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 4df2078e-d7ce-4a3f-904b-f3303071ffff] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1068.023647] env[62569]: INFO nova.scheduler.client.report [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Deleted allocation for migration de882a97-fdf6-4238-9efb-f038b01e1cf6 [ 1068.114999] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "ba73d02c-4c4f-441f-8142-f00af4524758" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.115272] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "ba73d02c-4c4f-441f-8142-f00af4524758" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.115480] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "ba73d02c-4c4f-441f-8142-f00af4524758-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.115665] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "ba73d02c-4c4f-441f-8142-f00af4524758-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.115841] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "ba73d02c-4c4f-441f-8142-f00af4524758-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.117827] env[62569]: INFO nova.compute.manager [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Terminating instance [ 1068.306372] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 2ce7e7ab-8a07-468f-9f9d-643d814981bd] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1068.529600] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 8.300s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.621124] env[62569]: DEBUG nova.compute.manager [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1068.621377] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1068.622302] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d241c454-3fd3-4f2b-980f-e9bb27b92d0c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.629815] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1068.630077] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e7e1bc27-32c3-4dcc-9223-e20a12c9c0b0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.694540] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1068.694767] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1068.694958] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Deleting the datastore file [datastore1] ba73d02c-4c4f-441f-8142-f00af4524758 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1068.695250] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-35ccb3d1-faac-456c-be85-5a33137e74b2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.701470] env[62569]: DEBUG oslo_vmware.api [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1068.701470] env[62569]: value = "task-1250697" [ 1068.701470] env[62569]: _type = "Task" [ 1068.701470] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.709911] env[62569]: DEBUG oslo_vmware.api [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250697, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.747655] env[62569]: DEBUG nova.network.neutron [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Port 8b78baca-9288-4ab0-b21d-7fb96baeaf42 binding to destination host cpu-1 is already ACTIVE {{(pid=62569) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1068.808963] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 9b5a1451-af46-4cd8-8e7a-f4560d8499e6] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1069.210900] env[62569]: DEBUG oslo_vmware.api [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250697, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131522} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.211135] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1069.211333] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1069.211525] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1069.211704] env[62569]: INFO nova.compute.manager [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1069.211939] env[62569]: DEBUG oslo.service.loopingcall [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1069.212150] env[62569]: DEBUG nova.compute.manager [-] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1069.212249] env[62569]: DEBUG nova.network.neutron [-] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1069.311954] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: d56faf39-6caf-4489-98d7-342fc5fb7d40] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1069.507880] env[62569]: DEBUG nova.compute.manager [req-9bf80273-1fe8-4399-b2f3-ef18a6c33e38 req-3bd1af1b-16a9-4f57-9129-721cd64418ab service nova] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Received event network-vif-deleted-e47e027b-ae9f-4e3d-a913-21e9088d31f0 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1069.507957] env[62569]: INFO nova.compute.manager [req-9bf80273-1fe8-4399-b2f3-ef18a6c33e38 req-3bd1af1b-16a9-4f57-9129-721cd64418ab service nova] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Neutron deleted interface e47e027b-ae9f-4e3d-a913-21e9088d31f0; detaching it from the instance and deleting it from the info cache [ 1069.508119] env[62569]: DEBUG nova.network.neutron [req-9bf80273-1fe8-4399-b2f3-ef18a6c33e38 req-3bd1af1b-16a9-4f57-9129-721cd64418ab service nova] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.765644] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.765892] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.766116] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.766308] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.766480] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.773791] env[62569]: INFO nova.compute.manager [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Terminating instance [ 1069.775103] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "659cc0b8-0f8e-4aa3-83ca-d8c04f355d81-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.775316] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "659cc0b8-0f8e-4aa3-83ca-d8c04f355d81-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.775489] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "659cc0b8-0f8e-4aa3-83ca-d8c04f355d81-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.814851] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 35350164-e295-40ea-85a6-e8e79c630e0f] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1069.985947] env[62569]: DEBUG nova.network.neutron [-] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.010427] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-46966add-9ddf-4d57-8b60-0c163bb1804d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.020614] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7138c95b-a648-44d7-8ccb-480a6a18d397 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.049143] env[62569]: DEBUG nova.compute.manager [req-9bf80273-1fe8-4399-b2f3-ef18a6c33e38 req-3bd1af1b-16a9-4f57-9129-721cd64418ab service nova] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Detach interface failed, port_id=e47e027b-ae9f-4e3d-a913-21e9088d31f0, reason: Instance ba73d02c-4c4f-441f-8142-f00af4524758 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1070.280086] env[62569]: DEBUG nova.compute.manager [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1070.280303] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1070.281342] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7521e544-1477-4b22-bc9c-782731ad402a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.288854] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1070.289669] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f292ebb-c96a-4d4d-9709-e7aad5f409b2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.295551] env[62569]: DEBUG oslo_vmware.api [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1070.295551] env[62569]: value = "task-1250698" [ 1070.295551] env[62569]: _type = "Task" [ 1070.295551] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.302695] env[62569]: DEBUG oslo_vmware.api [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250698, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.317552] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: bcc7f73b-083b-477e-bceb-4f984a95d219] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1070.488747] env[62569]: INFO nova.compute.manager [-] [instance: ba73d02c-4c4f-441f-8142-f00af4524758] Took 1.28 seconds to deallocate network for instance. [ 1070.805932] env[62569]: DEBUG oslo_vmware.api [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250698, 'name': PowerOffVM_Task, 'duration_secs': 0.157049} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.806234] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1070.806415] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1070.806679] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1512a65b-fe27-4fb9-af54-06ca8c66691e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.820837] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: abf56021-ddf8-4677-9e90-46dc96e89133] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1070.826022] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "refresh_cache-659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.826022] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired lock "refresh_cache-659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.826022] env[62569]: DEBUG nova.network.neutron [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1070.994662] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.994987] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.995241] env[62569]: DEBUG nova.objects.instance [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lazy-loading 'resources' on Instance uuid ba73d02c-4c4f-441f-8142-f00af4524758 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.326140] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 93eb1a09-269a-44c8-8f9d-3a2b7e0d715e] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1071.546495] env[62569]: DEBUG nova.network.neutron [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Updating instance_info_cache with network_info: [{"id": "8b78baca-9288-4ab0-b21d-7fb96baeaf42", "address": "fa:16:3e:7f:fc:2d", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b78baca-92", "ovs_interfaceid": "8b78baca-9288-4ab0-b21d-7fb96baeaf42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.607089] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96740a7e-8333-4834-b0f6-293895b95477 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.615238] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46495943-bc36-4913-b023-87079586939f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.647537] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-982fa194-b3a0-4093-a63f-ade1adda01bf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.655574] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d43acd4-cfcb-4774-87a8-03045cd1758f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.669430] env[62569]: DEBUG nova.compute.provider_tree [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1071.831021] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: d1fdff70-0401-49ca-bbbb-ef5eda266c9c] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1072.051044] env[62569]: DEBUG oslo_concurrency.lockutils [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Releasing lock "refresh_cache-659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.172065] env[62569]: DEBUG nova.scheduler.client.report [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1072.334089] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 09eeb937-aabc-418d-9fd3-3c7a72f4ef75] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1072.559331] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63c5ced3-d1e3-48c0-bc4b-7cc4db69482b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.566689] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c55d58e4-e15f-415f-805a-77ee198bd790 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.677423] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.682s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.697186] env[62569]: INFO nova.scheduler.client.report [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Deleted allocations for instance ba73d02c-4c4f-441f-8142-f00af4524758 [ 1072.837289] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 98089a34-074b-4bdb-92ae-f9e23d2551f1] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1072.912840] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1072.913040] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1072.913238] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Deleting the datastore file [datastore1] f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1072.913512] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-26d7725c-2d49-4d10-89de-4f9a45d823af {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.919170] env[62569]: DEBUG oslo_vmware.api [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for the task: (returnval){ [ 1072.919170] env[62569]: value = "task-1250700" [ 1072.919170] env[62569]: _type = "Task" [ 1072.919170] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.927038] env[62569]: DEBUG oslo_vmware.api [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250700, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.210977] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0cbf3176-a8ec-4c64-9338-be5849bea4f1 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "ba73d02c-4c4f-441f-8142-f00af4524758" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.096s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.340515] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: f751dfdb-e950-4821-b70d-4db78e03b9b7] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1073.430681] env[62569]: DEBUG oslo_vmware.api [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Task: {'id': task-1250700, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147565} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.430968] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1073.431228] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1073.431457] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1073.431695] env[62569]: INFO nova.compute.manager [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Took 3.15 seconds to destroy the instance on the hypervisor. [ 1073.431996] env[62569]: DEBUG oslo.service.loopingcall [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1073.432243] env[62569]: DEBUG nova.compute.manager [-] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1073.432350] env[62569]: DEBUG nova.network.neutron [-] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1073.675511] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23f09151-f29e-4da9-bf07-a9898191d269 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.698550] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46217f5a-48ce-4314-be9c-9d8a8392ea15 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.706190] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Updating instance '659cc0b8-0f8e-4aa3-83ca-d8c04f355d81' progress to 83 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1073.844649] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 6ccb78f3-d7f9-4090-9641-4633904efab8] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1073.853773] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "534c9ece-c8df-4528-83d8-c6c0ef8c4793" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.855047] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "534c9ece-c8df-4528-83d8-c6c0ef8c4793" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.855047] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "534c9ece-c8df-4528-83d8-c6c0ef8c4793-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.855047] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "534c9ece-c8df-4528-83d8-c6c0ef8c4793-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.855298] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "534c9ece-c8df-4528-83d8-c6c0ef8c4793-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.858021] env[62569]: INFO nova.compute.manager [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Terminating instance [ 1073.962497] env[62569]: DEBUG nova.compute.manager [req-b3e52d54-1780-44a0-b77a-57000e0c306b req-297390ac-729c-41df-9659-3003347020e5 service nova] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Received event network-vif-deleted-04dded5b-c41d-4dd5-b9bb-354115b85a5d {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1073.962497] env[62569]: INFO nova.compute.manager [req-b3e52d54-1780-44a0-b77a-57000e0c306b req-297390ac-729c-41df-9659-3003347020e5 service nova] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Neutron deleted interface 04dded5b-c41d-4dd5-b9bb-354115b85a5d; detaching it from the instance and deleting it from the info cache [ 1073.962497] env[62569]: DEBUG nova.network.neutron [req-b3e52d54-1780-44a0-b77a-57000e0c306b req-297390ac-729c-41df-9659-3003347020e5 service nova] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.214212] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1074.214212] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3eec2ca5-a1f5-44b1-a458-f7411294f029 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.222487] env[62569]: DEBUG oslo_vmware.api [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1074.222487] env[62569]: value = "task-1250701" [ 1074.222487] env[62569]: _type = "Task" [ 1074.222487] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.231209] env[62569]: DEBUG oslo_vmware.api [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250701, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.349105] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 188cac03-4034-4a02-973a-fb1906399fd1] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1074.362455] env[62569]: DEBUG nova.compute.manager [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1074.362677] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1074.363829] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9185019-565f-46a3-a08a-9ce7f31fce39 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.371449] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1074.371735] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-39ba7424-a121-4c69-ba65-9094512d619d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.378504] env[62569]: DEBUG oslo_vmware.api [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1074.378504] env[62569]: value = "task-1250702" [ 1074.378504] env[62569]: _type = "Task" [ 1074.378504] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.388636] env[62569]: DEBUG oslo_vmware.api [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250702, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.439299] env[62569]: DEBUG nova.network.neutron [-] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.466059] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f8c48334-2c43-447a-8dea-e47820467a67 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.475289] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89480538-dd58-4a0c-8d98-017d1e7bcf5b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.502138] env[62569]: DEBUG nova.compute.manager [req-b3e52d54-1780-44a0-b77a-57000e0c306b req-297390ac-729c-41df-9659-3003347020e5 service nova] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Detach interface failed, port_id=04dded5b-c41d-4dd5-b9bb-354115b85a5d, reason: Instance f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1074.733488] env[62569]: DEBUG oslo_vmware.api [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250701, 'name': PowerOnVM_Task, 'duration_secs': 0.463301} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.733769] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1074.733962] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-5f6b3e5e-f541-4ce5-8b47-1c8244453380 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Updating instance '659cc0b8-0f8e-4aa3-83ca-d8c04f355d81' progress to 100 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1074.853125] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: bc9e8bf5-ebdf-4d8c-ba26-cf449a68eb2a] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1074.888585] env[62569]: DEBUG oslo_vmware.api [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250702, 'name': PowerOffVM_Task, 'duration_secs': 0.221343} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.888867] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1074.889055] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1074.889308] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3d23d972-1f3f-4195-bf66-187bb7de2a6a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.941525] env[62569]: INFO nova.compute.manager [-] [instance: f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232] Took 1.51 seconds to deallocate network for instance. [ 1074.981584] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1074.981818] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1074.982016] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Deleting the datastore file [datastore1] 534c9ece-c8df-4528-83d8-c6c0ef8c4793 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1074.982308] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1efe05bd-b328-4ef1-915c-26998217b6fe {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.989396] env[62569]: DEBUG oslo_vmware.api [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for the task: (returnval){ [ 1074.989396] env[62569]: value = "task-1250704" [ 1074.989396] env[62569]: _type = "Task" [ 1074.989396] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.997213] env[62569]: DEBUG oslo_vmware.api [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250704, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.356105] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: a6ee3547-3a3d-4480-b451-d1fddd829b39] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1075.448647] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.448875] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.449077] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.470783] env[62569]: INFO nova.scheduler.client.report [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Deleted allocations for instance f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232 [ 1075.500084] env[62569]: DEBUG oslo_vmware.api [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Task: {'id': task-1250704, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169237} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.500380] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1075.500573] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1075.500757] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1075.500937] env[62569]: INFO nova.compute.manager [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1075.501199] env[62569]: DEBUG oslo.service.loopingcall [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1075.501399] env[62569]: DEBUG nova.compute.manager [-] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1075.501498] env[62569]: DEBUG nova.network.neutron [-] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1075.859299] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 9c1c0379-3968-464b-b587-8fac704404d5] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1075.978825] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cb4f8a40-209a-489c-b2de-892c2001585f tempest-DeleteServersTestJSON-1073539779 tempest-DeleteServersTestJSON-1073539779-project-member] Lock "f2a1d6e4-39b5-4f08-89fd-1cfa1eadb232" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.213s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.988365] env[62569]: DEBUG nova.compute.manager [req-585b585d-3f2b-4f32-8431-ec99df522ff7 req-3f3e2d98-0c63-46c0-86f0-01c10643ca00 service nova] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Received event network-vif-deleted-37c7a902-07de-45dc-b683-311b84a5edc5 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1075.988365] env[62569]: INFO nova.compute.manager [req-585b585d-3f2b-4f32-8431-ec99df522ff7 req-3f3e2d98-0c63-46c0-86f0-01c10643ca00 service nova] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Neutron deleted interface 37c7a902-07de-45dc-b683-311b84a5edc5; detaching it from the instance and deleting it from the info cache [ 1075.988365] env[62569]: DEBUG nova.network.neutron [req-585b585d-3f2b-4f32-8431-ec99df522ff7 req-3f3e2d98-0c63-46c0-86f0-01c10643ca00 service nova] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.214033] env[62569]: DEBUG nova.network.neutron [-] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.362083] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 108d5bde-e463-4a69-816b-bdd3c03e13cd] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1076.491056] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f31629b2-e755-412f-9a58-6d63c60a1e18 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.499697] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1781b63-86bc-4ccd-a335-58cb3022842e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.522863] env[62569]: DEBUG nova.compute.manager [req-585b585d-3f2b-4f32-8431-ec99df522ff7 req-3f3e2d98-0c63-46c0-86f0-01c10643ca00 service nova] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Detach interface failed, port_id=37c7a902-07de-45dc-b683-311b84a5edc5, reason: Instance 534c9ece-c8df-4528-83d8-c6c0ef8c4793 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1076.719688] env[62569]: INFO nova.compute.manager [-] [instance: 534c9ece-c8df-4528-83d8-c6c0ef8c4793] Took 1.22 seconds to deallocate network for instance. [ 1076.865024] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 298efba9-226d-4105-8b32-76b3dddddb0b] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1077.192586] env[62569]: DEBUG nova.network.neutron [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Port 8b78baca-9288-4ab0-b21d-7fb96baeaf42 binding to destination host cpu-1 is already ACTIVE {{(pid=62569) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1077.192908] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "refresh_cache-659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.193141] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired lock "refresh_cache-659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.193326] env[62569]: DEBUG nova.network.neutron [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1077.227409] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.227708] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.228092] env[62569]: DEBUG nova.objects.instance [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lazy-loading 'resources' on Instance uuid 534c9ece-c8df-4528-83d8-c6c0ef8c4793 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1077.368741] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 691d8fbe-b9ee-454f-bd7b-14520e53ed26] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1077.804942] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c50ad97-0c7a-438e-87d2-c936b308270a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.812396] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a426c242-bd8b-4d3d-a444-12662d689835 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.844977] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53e3b80a-7582-46ac-bb47-7ccc3a9ad05e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.852996] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b83f12ae-e9e2-4450-bae8-c161e2987865 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.865843] env[62569]: DEBUG nova.compute.provider_tree [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1077.872714] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 31ce29fa-4f60-4404-b830-21ad196f78b5] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1077.999321] env[62569]: DEBUG nova.network.neutron [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Updating instance_info_cache with network_info: [{"id": "8b78baca-9288-4ab0-b21d-7fb96baeaf42", "address": "fa:16:3e:7f:fc:2d", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b78baca-92", "ovs_interfaceid": "8b78baca-9288-4ab0-b21d-7fb96baeaf42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.371030] env[62569]: DEBUG nova.scheduler.client.report [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1078.376588] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: c8edf395-ea21-4ba8-991e-0f43fe4ee830] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1078.502150] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Releasing lock "refresh_cache-659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.874734] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.647s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.879179] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 05db90a6-3faf-4878-b782-ed17df47ed5f] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1078.892895] env[62569]: INFO nova.scheduler.client.report [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Deleted allocations for instance 534c9ece-c8df-4528-83d8-c6c0ef8c4793 [ 1079.005995] env[62569]: DEBUG nova.compute.manager [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62569) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:900}} [ 1079.382356] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: a5b3c345-149b-4dc3-84ff-fe72183efadb] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1079.400469] env[62569]: DEBUG oslo_concurrency.lockutils [None req-de84edb2-83f4-4a0c-9eb1-250cfdd71ba6 tempest-ServersTestJSON-1198898795 tempest-ServersTestJSON-1198898795-project-member] Lock "534c9ece-c8df-4528-83d8-c6c0ef8c4793" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.546s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.885845] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: c9568fcb-745f-4c04-8487-a38a617aab7c] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1080.111355] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.111568] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.389074] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 6e83fd32-215a-42e4-99e7-72bdce25555e] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1080.616080] env[62569]: DEBUG nova.objects.instance [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lazy-loading 'migration_context' on Instance uuid 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1080.892569] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: e4eadcdf-a04a-4255-ba1c-fe20156c655f] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1081.196951] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aebd3621-313b-46b3-a594-4b668910d7e0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.206845] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fea26f6a-0754-4c37-ab0d-d34695499b5f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.243111] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91ba29d6-7b83-447c-8668-2ee683b265b1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.250762] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d838edf-b962-4a57-81b4-a9750661d2fc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.264967] env[62569]: DEBUG nova.compute.provider_tree [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1081.396136] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: cb9d8991-b7ba-4bf7-9d22-1391f4cea768] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1081.772018] env[62569]: DEBUG nova.scheduler.client.report [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1081.899579] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: b47cd2d7-0cd2-41af-8ed1-a6dfca323516] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1082.403165] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: fdeac8b8-a34d-4b4c-8f72-281f84e22ebe] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1082.783609] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.672s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.907144] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 390bcf25-689d-46ad-bffb-3670c3729397] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1083.410071] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 807d3025-d6a7-4778-a829-a61e2c7495c9] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1083.652908] env[62569]: DEBUG nova.compute.manager [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Stashing vm_state: active {{(pid=62569) _prep_resize /opt/stack/nova/nova/compute/manager.py:5997}} [ 1083.913501] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: cefbca2e-609d-4954-bec6-52ffe095446f] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1084.176021] env[62569]: DEBUG oslo_concurrency.lockutils [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.176021] env[62569]: DEBUG oslo_concurrency.lockutils [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.316480] env[62569]: INFO nova.compute.manager [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Swapping old allocation on dict_keys(['fa06556a-5785-4014-b8bd-bc240a0cf716']) held by migration 56485982-7035-45de-a101-71589195430f for instance [ 1084.339080] env[62569]: DEBUG nova.scheduler.client.report [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Overwriting current allocation {'allocations': {'fa06556a-5785-4014-b8bd-bc240a0cf716': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 156}}, 'project_id': '00bdb4c3f88f4d61834f7e961629ef3a', 'user_id': 'b9e0da546a03426782fcacd06a7afa8a', 'consumer_generation': 1} on consumer 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81 {{(pid=62569) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 1084.416479] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "refresh_cache-659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.416749] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired lock "refresh_cache-659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.416850] env[62569]: DEBUG nova.network.neutron [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1084.418227] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: eb70341b-4282-4eca-b6a2-374db7c521c5] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1084.679773] env[62569]: INFO nova.compute.claims [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1084.923092] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 337d88a2-30b9-4846-929e-042bd7a64a65] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1085.131853] env[62569]: DEBUG nova.network.neutron [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Updating instance_info_cache with network_info: [{"id": "8b78baca-9288-4ab0-b21d-7fb96baeaf42", "address": "fa:16:3e:7f:fc:2d", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b78baca-92", "ovs_interfaceid": "8b78baca-9288-4ab0-b21d-7fb96baeaf42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.186206] env[62569]: INFO nova.compute.resource_tracker [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Updating resource usage from migration a3c050a1-397b-420f-b575-14e1ae1b7da5 [ 1085.240486] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f183ac93-49ac-4a85-82dd-c9d474e65c60 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.248090] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15b6c5da-25ca-44a4-b178-57350c1c2b1d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.277807] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c78f3a-5816-47f8-b211-73cd0522ccd5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.284529] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c12919e-2a77-4981-894d-9e7308ded79e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.297339] env[62569]: DEBUG nova.compute.provider_tree [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1085.426251] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: b3ccc28b-68d0-461d-a67d-b5ad179a80f9] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1085.634503] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Releasing lock "refresh_cache-659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.635519] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48e56f99-c980-49e1-982d-e2670d6a5ec8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.642926] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3744a732-cb9a-47df-b1e8-92102bf5e6c3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.800789] env[62569]: DEBUG nova.scheduler.client.report [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1085.929390] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 06067572-a9fd-43a4-91cd-383a62f53885] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1086.305580] env[62569]: DEBUG oslo_concurrency.lockutils [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.130s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.305790] env[62569]: INFO nova.compute.manager [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Migrating [ 1086.434064] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 37a4c60a-e660-4c91-bcfe-72638a4397b0] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1086.722103] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1086.722520] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-112e034e-502e-4fe0-90a4-3305f1cfc3c5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.732109] env[62569]: DEBUG oslo_vmware.api [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1086.732109] env[62569]: value = "task-1250706" [ 1086.732109] env[62569]: _type = "Task" [ 1086.732109] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.739729] env[62569]: DEBUG oslo_vmware.api [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250706, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.819826] env[62569]: DEBUG oslo_concurrency.lockutils [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "refresh_cache-96167629-8777-4a1c-b564-c70d1bb59b90" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.820089] env[62569]: DEBUG oslo_concurrency.lockutils [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "refresh_cache-96167629-8777-4a1c-b564-c70d1bb59b90" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.820304] env[62569]: DEBUG nova.network.neutron [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1086.936489] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: b6e5eefc-8c06-445b-a3af-9404578b6179] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1087.241804] env[62569]: DEBUG oslo_vmware.api [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250706, 'name': PowerOffVM_Task, 'duration_secs': 0.211431} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.242044] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1087.242693] env[62569]: DEBUG nova.virt.hardware [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1087.242912] env[62569]: DEBUG nova.virt.hardware [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1087.243086] env[62569]: DEBUG nova.virt.hardware [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1087.243311] env[62569]: DEBUG nova.virt.hardware [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1087.243472] env[62569]: DEBUG nova.virt.hardware [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1087.243626] env[62569]: DEBUG nova.virt.hardware [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1087.243837] env[62569]: DEBUG nova.virt.hardware [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1087.244009] env[62569]: DEBUG nova.virt.hardware [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1087.244193] env[62569]: DEBUG nova.virt.hardware [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1087.244385] env[62569]: DEBUG nova.virt.hardware [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1087.244569] env[62569]: DEBUG nova.virt.hardware [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1087.249376] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-618fecc7-e6d0-42b8-80aa-407e0e87ae67 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.264755] env[62569]: DEBUG oslo_vmware.api [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1087.264755] env[62569]: value = "task-1250707" [ 1087.264755] env[62569]: _type = "Task" [ 1087.264755] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.271815] env[62569]: DEBUG oslo_vmware.api [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250707, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.440258] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 0b2e5755-6d3e-4b03-9f05-aa2bc0b6e010] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1087.533207] env[62569]: DEBUG nova.network.neutron [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Updating instance_info_cache with network_info: [{"id": "d5f53bf1-303b-435e-941a-47c7ab293484", "address": "fa:16:3e:e8:8d:ca", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5f53bf1-30", "ovs_interfaceid": "d5f53bf1-303b-435e-941a-47c7ab293484", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.775330] env[62569]: DEBUG oslo_vmware.api [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250707, 'name': ReconfigVM_Task, 'duration_secs': 0.151025} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.776108] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf029e48-a338-42f0-b430-b7a12f493f5a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.796154] env[62569]: DEBUG nova.virt.hardware [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1087.796410] env[62569]: DEBUG nova.virt.hardware [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1087.796608] env[62569]: DEBUG nova.virt.hardware [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1087.796805] env[62569]: DEBUG nova.virt.hardware [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1087.796961] env[62569]: DEBUG nova.virt.hardware [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1087.797134] env[62569]: DEBUG nova.virt.hardware [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1087.797420] env[62569]: DEBUG nova.virt.hardware [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1087.797541] env[62569]: DEBUG nova.virt.hardware [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1087.797694] env[62569]: DEBUG nova.virt.hardware [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1087.797862] env[62569]: DEBUG nova.virt.hardware [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1087.798052] env[62569]: DEBUG nova.virt.hardware [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1087.798813] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32e870b0-0066-4a03-896f-5419b270d7e8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.804168] env[62569]: DEBUG oslo_vmware.api [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1087.804168] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]524355fc-a2fc-81d1-611f-1771b0b0e44a" [ 1087.804168] env[62569]: _type = "Task" [ 1087.804168] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.811311] env[62569]: DEBUG oslo_vmware.api [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]524355fc-a2fc-81d1-611f-1771b0b0e44a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.943127] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 4feb7b78-9f7c-4e64-b0a7-870ed73adf97] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1088.036281] env[62569]: DEBUG oslo_concurrency.lockutils [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "refresh_cache-96167629-8777-4a1c-b564-c70d1bb59b90" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.314206] env[62569]: DEBUG oslo_vmware.api [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]524355fc-a2fc-81d1-611f-1771b0b0e44a, 'name': SearchDatastore_Task, 'duration_secs': 0.007196} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.319434] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Reconfiguring VM instance instance-00000066 to detach disk 2000 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1088.319702] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-443f989c-e890-493b-96bf-b25921e9ce24 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.336630] env[62569]: DEBUG oslo_vmware.api [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1088.336630] env[62569]: value = "task-1250708" [ 1088.336630] env[62569]: _type = "Task" [ 1088.336630] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.345136] env[62569]: DEBUG oslo_vmware.api [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250708, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.446348] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 22094c32-5f50-4f86-a77b-cd4adcf8998a] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1088.847143] env[62569]: DEBUG oslo_vmware.api [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250708, 'name': ReconfigVM_Task, 'duration_secs': 0.211589} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.847420] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Reconfigured VM instance instance-00000066 to detach disk 2000 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1088.848160] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c72ed3f5-4b97-47e2-b69c-acaa37fe3717 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.871777] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81/659cc0b8-0f8e-4aa3-83ca-d8c04f355d81.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1088.872020] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-668db171-478b-4778-a68a-4946e0f6575d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.889072] env[62569]: DEBUG oslo_vmware.api [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1088.889072] env[62569]: value = "task-1250709" [ 1088.889072] env[62569]: _type = "Task" [ 1088.889072] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.897254] env[62569]: DEBUG oslo_vmware.api [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250709, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.949985] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 77a1b192-6aff-4fee-93d7-57cebcdce626] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1089.399411] env[62569]: DEBUG oslo_vmware.api [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250709, 'name': ReconfigVM_Task, 'duration_secs': 0.281798} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.399665] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81/659cc0b8-0f8e-4aa3-83ca-d8c04f355d81.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1089.400518] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42cdfcf0-53fd-466f-9d5a-b756d576687f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.420097] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad655e0b-c825-4db8-8c00-3c8ae77a08c1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.439636] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a37ebd-d942-439c-a8b1-166de4279137 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.460160] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: fd763e12-5b29-48f8-8256-cb6205e0d119] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1089.462526] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e11e5b6c-9809-425d-beaf-17899809ade6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.468871] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1089.469116] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-30f2e944-894b-424c-86ec-6d0f5f3942d5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.474509] env[62569]: DEBUG oslo_vmware.api [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1089.474509] env[62569]: value = "task-1250710" [ 1089.474509] env[62569]: _type = "Task" [ 1089.474509] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.481364] env[62569]: DEBUG oslo_vmware.api [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250710, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.551644] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30172fa3-f453-472f-944c-d6670f5f2788 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.569864] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Updating instance '96167629-8777-4a1c-b564-c70d1bb59b90' progress to 0 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1089.966072] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: c515d85e-bcb5-4bac-bacb-1e558f38171f] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1089.984570] env[62569]: DEBUG oslo_vmware.api [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250710, 'name': PowerOnVM_Task, 'duration_secs': 0.358801} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.984735] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1090.075559] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1090.075847] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e8580a4-d80b-4a51-a2da-6bbf0e9893d5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.082683] env[62569]: DEBUG oslo_vmware.api [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1090.082683] env[62569]: value = "task-1250711" [ 1090.082683] env[62569]: _type = "Task" [ 1090.082683] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.090874] env[62569]: DEBUG oslo_vmware.api [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250711, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.469700] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 015140aa-8336-40e2-856c-723277a48f6e] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1090.592207] env[62569]: DEBUG oslo_vmware.api [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250711, 'name': PowerOffVM_Task, 'duration_secs': 0.174394} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.592506] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1090.592704] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Updating instance '96167629-8777-4a1c-b564-c70d1bb59b90' progress to 17 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1090.973050] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 94392d48-223d-4205-9627-25488a468769] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1091.037052] env[62569]: INFO nova.compute.manager [None req-4b745074-1e1f-417a-87f0-b6390380b6dd tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Updating instance to original state: 'active' [ 1091.098547] env[62569]: DEBUG nova.virt.hardware [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1091.098798] env[62569]: DEBUG nova.virt.hardware [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1091.098961] env[62569]: DEBUG nova.virt.hardware [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1091.099165] env[62569]: DEBUG nova.virt.hardware [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1091.099317] env[62569]: DEBUG nova.virt.hardware [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1091.099469] env[62569]: DEBUG nova.virt.hardware [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1091.099675] env[62569]: DEBUG nova.virt.hardware [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1091.099839] env[62569]: DEBUG nova.virt.hardware [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1091.100020] env[62569]: DEBUG nova.virt.hardware [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1091.100197] env[62569]: DEBUG nova.virt.hardware [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1091.100374] env[62569]: DEBUG nova.virt.hardware [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1091.105579] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c8d4a20-89f9-4432-901d-c15e7670f220 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.121210] env[62569]: DEBUG oslo_vmware.api [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1091.121210] env[62569]: value = "task-1250712" [ 1091.121210] env[62569]: _type = "Task" [ 1091.121210] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.129092] env[62569]: DEBUG oslo_vmware.api [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250712, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.476339] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: a0b406c3-9466-41bd-9de1-e675cab2ceef] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1091.631620] env[62569]: DEBUG oslo_vmware.api [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250712, 'name': ReconfigVM_Task, 'duration_secs': 0.159008} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.631937] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Updating instance '96167629-8777-4a1c-b564-c70d1bb59b90' progress to 33 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1091.979986] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 3f504d5c-3431-4a80-a7d9-e52af01cf24e] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1092.114971] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.115336] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.115749] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "659cc0b8-0f8e-4aa3-83ca-d8c04f355d81-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.116329] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "659cc0b8-0f8e-4aa3-83ca-d8c04f355d81-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.116579] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "659cc0b8-0f8e-4aa3-83ca-d8c04f355d81-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.118989] env[62569]: INFO nova.compute.manager [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Terminating instance [ 1092.138393] env[62569]: DEBUG nova.virt.hardware [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1092.138636] env[62569]: DEBUG nova.virt.hardware [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1092.138812] env[62569]: DEBUG nova.virt.hardware [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1092.139036] env[62569]: DEBUG nova.virt.hardware [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1092.139214] env[62569]: DEBUG nova.virt.hardware [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1092.139381] env[62569]: DEBUG nova.virt.hardware [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1092.139607] env[62569]: DEBUG nova.virt.hardware [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1092.139788] env[62569]: DEBUG nova.virt.hardware [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1092.139976] env[62569]: DEBUG nova.virt.hardware [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1092.140175] env[62569]: DEBUG nova.virt.hardware [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1092.140355] env[62569]: DEBUG nova.virt.hardware [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1092.145757] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1092.146261] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f7417ab-cfcf-4f87-b666-a74e2601e0e4 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.164654] env[62569]: DEBUG oslo_vmware.api [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1092.164654] env[62569]: value = "task-1250713" [ 1092.164654] env[62569]: _type = "Task" [ 1092.164654] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.172406] env[62569]: DEBUG oslo_vmware.api [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250713, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.484049] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 93c8445d-689c-4f77-836f-95eafafbb6a2] Instance has had 0 of 5 cleanup attempts {{(pid=62569) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11621}} [ 1092.623833] env[62569]: DEBUG nova.compute.manager [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1092.623833] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1092.624049] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d200d535-3699-4c3c-a381-eb805d7a2d9f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.631188] env[62569]: DEBUG oslo_vmware.api [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1092.631188] env[62569]: value = "task-1250714" [ 1092.631188] env[62569]: _type = "Task" [ 1092.631188] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.639788] env[62569]: DEBUG oslo_vmware.api [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250714, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.673519] env[62569]: DEBUG oslo_vmware.api [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250713, 'name': ReconfigVM_Task, 'duration_secs': 0.206221} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.673820] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1092.674646] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18c97f3d-fb0e-4771-8225-ba9d62e6c268 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.698175] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] 96167629-8777-4a1c-b564-c70d1bb59b90/96167629-8777-4a1c-b564-c70d1bb59b90.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1092.698489] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f70f8bc3-8aa9-4062-b5b6-704370d708bc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.716456] env[62569]: DEBUG oslo_vmware.api [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1092.716456] env[62569]: value = "task-1250715" [ 1092.716456] env[62569]: _type = "Task" [ 1092.716456] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.724077] env[62569]: DEBUG oslo_vmware.api [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250715, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.988064] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1093.141216] env[62569]: DEBUG oslo_vmware.api [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250714, 'name': PowerOffVM_Task, 'duration_secs': 0.210089} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.141493] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1093.141696] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Volume detach. Driver type: vmdk {{(pid=62569) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1093.141895] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269507', 'volume_id': 'aa9283c3-c013-498f-8986-2367724e045f', 'name': 'volume-aa9283c3-c013-498f-8986-2367724e045f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '659cc0b8-0f8e-4aa3-83ca-d8c04f355d81', 'attached_at': '2024-10-10T15:41:34.000000', 'detached_at': '', 'volume_id': 'aa9283c3-c013-498f-8986-2367724e045f', 'serial': 'aa9283c3-c013-498f-8986-2367724e045f'} {{(pid=62569) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1093.142673] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0a24b1c-4832-4529-adf7-1526144f0cea {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.162740] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9840bf7-9827-4778-a083-1ea3abaf86fd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.168871] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8051e8ac-c9b5-40fe-a827-4965cb05ecfc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.188599] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e45a4c40-3bfa-4d2f-9e81-ec5b413085e5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.202433] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] The volume has not been displaced from its original location: [datastore1] volume-aa9283c3-c013-498f-8986-2367724e045f/volume-aa9283c3-c013-498f-8986-2367724e045f.vmdk. No consolidation needed. {{(pid=62569) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1093.207678] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Reconfiguring VM instance instance-00000066 to detach disk 2001 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1093.207930] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e22ca0f3-4c1e-493d-8914-d8eaa8215c88 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.227701] env[62569]: DEBUG oslo_vmware.api [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250715, 'name': ReconfigVM_Task, 'duration_secs': 0.25393} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.228753] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Reconfigured VM instance instance-0000006c to attach disk [datastore2] 96167629-8777-4a1c-b564-c70d1bb59b90/96167629-8777-4a1c-b564-c70d1bb59b90.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1093.229104] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Updating instance '96167629-8777-4a1c-b564-c70d1bb59b90' progress to 50 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1093.232438] env[62569]: DEBUG oslo_vmware.api [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1093.232438] env[62569]: value = "task-1250716" [ 1093.232438] env[62569]: _type = "Task" [ 1093.232438] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.241066] env[62569]: DEBUG oslo_vmware.api [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250716, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.737027] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a652af-2bc0-4476-8f6f-034f0e6a2ed1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.747799] env[62569]: DEBUG oslo_vmware.api [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250716, 'name': ReconfigVM_Task, 'duration_secs': 0.197502} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.760812] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Reconfigured VM instance instance-00000066 to detach disk 2001 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1093.765678] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c4107ada-1d60-45dc-9a1b-83dd7d580abf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.775667] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-516a557d-dbf0-4e62-8c25-e8cadce82694 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.792646] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Updating instance '96167629-8777-4a1c-b564-c70d1bb59b90' progress to 67 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1093.797096] env[62569]: DEBUG oslo_vmware.api [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1093.797096] env[62569]: value = "task-1250717" [ 1093.797096] env[62569]: _type = "Task" [ 1093.797096] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.805024] env[62569]: DEBUG oslo_vmware.api [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250717, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.311132] env[62569]: DEBUG oslo_vmware.api [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250717, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.336148] env[62569]: DEBUG nova.network.neutron [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Port d5f53bf1-303b-435e-941a-47c7ab293484 binding to destination host cpu-1 is already ACTIVE {{(pid=62569) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1094.812954] env[62569]: DEBUG oslo_vmware.api [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250717, 'name': ReconfigVM_Task, 'duration_secs': 0.796494} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.813393] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269507', 'volume_id': 'aa9283c3-c013-498f-8986-2367724e045f', 'name': 'volume-aa9283c3-c013-498f-8986-2367724e045f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '659cc0b8-0f8e-4aa3-83ca-d8c04f355d81', 'attached_at': '2024-10-10T15:41:34.000000', 'detached_at': '', 'volume_id': 'aa9283c3-c013-498f-8986-2367724e045f', 'serial': 'aa9283c3-c013-498f-8986-2367724e045f'} {{(pid=62569) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1094.813704] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1094.814524] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc309a01-c6ca-42b9-aa16-12967d1e8d2c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.820736] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1094.820960] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bea4f81d-70cd-4ade-a628-76f4b9e6ea57 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.880359] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1094.880691] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1094.880913] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Deleting the datastore file [datastore2] 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1094.881207] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-163e4440-9bab-4d2a-800c-9be68ead8362 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.889135] env[62569]: DEBUG oslo_vmware.api [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1094.889135] env[62569]: value = "task-1250719" [ 1094.889135] env[62569]: _type = "Task" [ 1094.889135] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.897397] env[62569]: DEBUG oslo_vmware.api [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250719, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.358825] env[62569]: DEBUG oslo_concurrency.lockutils [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "96167629-8777-4a1c-b564-c70d1bb59b90-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.359155] env[62569]: DEBUG oslo_concurrency.lockutils [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "96167629-8777-4a1c-b564-c70d1bb59b90-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.359356] env[62569]: DEBUG oslo_concurrency.lockutils [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "96167629-8777-4a1c-b564-c70d1bb59b90-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.399312] env[62569]: DEBUG oslo_vmware.api [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250719, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157169} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.399557] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1095.399752] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1095.399931] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1095.400126] env[62569]: INFO nova.compute.manager [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Took 2.78 seconds to destroy the instance on the hypervisor. [ 1095.400360] env[62569]: DEBUG oslo.service.loopingcall [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1095.400552] env[62569]: DEBUG nova.compute.manager [-] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1095.400645] env[62569]: DEBUG nova.network.neutron [-] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1096.049089] env[62569]: DEBUG nova.compute.manager [req-fee6c095-23ec-4cef-a941-7963b9486ca0 req-b7335af4-921c-4a5b-b770-62fc6ee11bba service nova] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Received event network-vif-deleted-8b78baca-9288-4ab0-b21d-7fb96baeaf42 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1096.049346] env[62569]: INFO nova.compute.manager [req-fee6c095-23ec-4cef-a941-7963b9486ca0 req-b7335af4-921c-4a5b-b770-62fc6ee11bba service nova] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Neutron deleted interface 8b78baca-9288-4ab0-b21d-7fb96baeaf42; detaching it from the instance and deleting it from the info cache [ 1096.049489] env[62569]: DEBUG nova.network.neutron [req-fee6c095-23ec-4cef-a941-7963b9486ca0 req-b7335af4-921c-4a5b-b770-62fc6ee11bba service nova] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.388857] env[62569]: DEBUG oslo_concurrency.lockutils [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "refresh_cache-96167629-8777-4a1c-b564-c70d1bb59b90" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.389135] env[62569]: DEBUG oslo_concurrency.lockutils [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "refresh_cache-96167629-8777-4a1c-b564-c70d1bb59b90" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.389331] env[62569]: DEBUG nova.network.neutron [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1096.520156] env[62569]: DEBUG nova.network.neutron [-] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.552455] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4f42b60c-9302-4d86-a234-43290b70508e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.562137] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-259d912a-2b64-40b5-bb32-85a0e111ac99 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.585877] env[62569]: DEBUG nova.compute.manager [req-fee6c095-23ec-4cef-a941-7963b9486ca0 req-b7335af4-921c-4a5b-b770-62fc6ee11bba service nova] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Detach interface failed, port_id=8b78baca-9288-4ab0-b21d-7fb96baeaf42, reason: Instance 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1097.022610] env[62569]: INFO nova.compute.manager [-] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Took 1.62 seconds to deallocate network for instance. [ 1097.142504] env[62569]: DEBUG nova.network.neutron [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Updating instance_info_cache with network_info: [{"id": "d5f53bf1-303b-435e-941a-47c7ab293484", "address": "fa:16:3e:e8:8d:ca", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5f53bf1-30", "ovs_interfaceid": "d5f53bf1-303b-435e-941a-47c7ab293484", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.566797] env[62569]: INFO nova.compute.manager [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81] Took 0.54 seconds to detach 1 volumes for instance. [ 1097.645940] env[62569]: DEBUG oslo_concurrency.lockutils [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "refresh_cache-96167629-8777-4a1c-b564-c70d1bb59b90" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1098.073683] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.073969] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.074226] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.096334] env[62569]: INFO nova.scheduler.client.report [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Deleted allocations for instance 659cc0b8-0f8e-4aa3-83ca-d8c04f355d81 [ 1098.166063] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63942312-f5cf-4b2e-88f8-1779c0634c4d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.185294] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fc158fc-e2e3-45a0-bb49-e61ffd293ad5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.192218] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Updating instance '96167629-8777-4a1c-b564-c70d1bb59b90' progress to 83 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1098.389206] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0db13343-9297-4770-894a-3a071a74d1ca tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "cf077eb8-8860-46f0-ae84-eed58c29873b" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.389464] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0db13343-9297-4770-894a-3a071a74d1ca tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cf077eb8-8860-46f0-ae84-eed58c29873b" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.605181] env[62569]: DEBUG oslo_concurrency.lockutils [None req-02723bef-0100-4d17-9ae4-7519fdcdca83 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "659cc0b8-0f8e-4aa3-83ca-d8c04f355d81" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.490s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.698916] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1098.699405] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-38846d74-ea96-44cc-a64e-a44a89c51361 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.706985] env[62569]: DEBUG oslo_vmware.api [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1098.706985] env[62569]: value = "task-1250720" [ 1098.706985] env[62569]: _type = "Task" [ 1098.706985] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.714745] env[62569]: DEBUG oslo_vmware.api [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250720, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.892569] env[62569]: DEBUG nova.compute.utils [None req-0db13343-9297-4770-894a-3a071a74d1ca tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1099.216778] env[62569]: DEBUG oslo_vmware.api [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250720, 'name': PowerOnVM_Task, 'duration_secs': 0.358719} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.217120] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1099.217289] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-43724702-3114-4511-9a45-def6f748a00a tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Updating instance '96167629-8777-4a1c-b564-c70d1bb59b90' progress to 100 {{(pid=62569) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1099.287168] env[62569]: DEBUG oslo_concurrency.lockutils [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "27b14750-e349-47af-affe-02d10050aad1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.287423] env[62569]: DEBUG oslo_concurrency.lockutils [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "27b14750-e349-47af-affe-02d10050aad1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.395982] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0db13343-9297-4770-894a-3a071a74d1ca tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cf077eb8-8860-46f0-ae84-eed58c29873b" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.791059] env[62569]: DEBUG nova.compute.manager [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1100.312106] env[62569]: DEBUG oslo_concurrency.lockutils [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.312374] env[62569]: DEBUG oslo_concurrency.lockutils [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.313880] env[62569]: INFO nova.compute.claims [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1100.483919] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1100.522809] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0db13343-9297-4770-894a-3a071a74d1ca tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "cf077eb8-8860-46f0-ae84-eed58c29873b" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.523194] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0db13343-9297-4770-894a-3a071a74d1ca tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cf077eb8-8860-46f0-ae84-eed58c29873b" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.523521] env[62569]: INFO nova.compute.manager [None req-0db13343-9297-4770-894a-3a071a74d1ca tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Attaching volume b29ad30c-88ee-4fac-a8b0-ecc6276571dd to /dev/sdb [ 1100.587855] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8faed614-15b5-45a7-bbec-926228dc344e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.596185] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba966828-312d-4bc2-8df4-19c3cf3c9275 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.610048] env[62569]: DEBUG nova.virt.block_device [None req-0db13343-9297-4770-894a-3a071a74d1ca tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Updating existing volume attachment record: fc4be6b4-cf04-466e-a7d3-77507be5e2ef {{(pid=62569) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1101.181183] env[62569]: DEBUG nova.network.neutron [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Port d5f53bf1-303b-435e-941a-47c7ab293484 binding to destination host cpu-1 is already ACTIVE {{(pid=62569) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1101.181521] env[62569]: DEBUG oslo_concurrency.lockutils [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "refresh_cache-96167629-8777-4a1c-b564-c70d1bb59b90" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1101.181689] env[62569]: DEBUG oslo_concurrency.lockutils [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "refresh_cache-96167629-8777-4a1c-b564-c70d1bb59b90" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.181861] env[62569]: DEBUG nova.network.neutron [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1101.382561] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24bf31f2-05fa-4445-ad56-e71b422063ff {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.390638] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d4a4cf-b462-422d-a971-34c8ad63055d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.419724] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a4d94b-877d-40c3-ae32-6e69389bef72 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.427231] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c27e13-8807-49d0-b0c9-c4b28d5a93a9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.440165] env[62569]: DEBUG nova.compute.provider_tree [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1101.938467] env[62569]: DEBUG nova.network.neutron [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Updating instance_info_cache with network_info: [{"id": "d5f53bf1-303b-435e-941a-47c7ab293484", "address": "fa:16:3e:e8:8d:ca", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5f53bf1-30", "ovs_interfaceid": "d5f53bf1-303b-435e-941a-47c7ab293484", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.943438] env[62569]: DEBUG nova.scheduler.client.report [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1102.441941] env[62569]: DEBUG oslo_concurrency.lockutils [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "refresh_cache-96167629-8777-4a1c-b564-c70d1bb59b90" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1102.449483] env[62569]: DEBUG oslo_concurrency.lockutils [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.137s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.449965] env[62569]: DEBUG nova.compute.manager [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1102.945958] env[62569]: DEBUG nova.compute.manager [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62569) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:900}} [ 1102.946265] env[62569]: DEBUG oslo_concurrency.lockutils [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.946575] env[62569]: DEBUG oslo_concurrency.lockutils [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.954991] env[62569]: DEBUG nova.compute.utils [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1102.956427] env[62569]: DEBUG nova.compute.manager [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1102.956637] env[62569]: DEBUG nova.network.neutron [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1102.997163] env[62569]: DEBUG nova.policy [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b9e0da546a03426782fcacd06a7afa8a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '00bdb4c3f88f4d61834f7e961629ef3a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 1103.258377] env[62569]: DEBUG nova.network.neutron [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Successfully created port: 26f272f3-4864-483c-a76f-3bcf9efa4ab8 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1103.449753] env[62569]: DEBUG nova.objects.instance [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lazy-loading 'migration_context' on Instance uuid 96167629-8777-4a1c-b564-c70d1bb59b90 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1103.461975] env[62569]: DEBUG nova.compute.manager [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1104.019690] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ced2dbf9-b337-43f5-97ac-6da8f0f8161e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.027168] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16ba6d2b-f564-478e-9820-6023c4fadf33 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.058935] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7968d336-e86a-485d-8619-c7f4c011e7f5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.066344] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec6c7278-4caf-4b92-97e5-a81e277fa2c8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.080578] env[62569]: DEBUG nova.compute.provider_tree [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1104.471807] env[62569]: DEBUG nova.compute.manager [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1104.495505] env[62569]: DEBUG nova.virt.hardware [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1104.496191] env[62569]: DEBUG nova.virt.hardware [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1104.496191] env[62569]: DEBUG nova.virt.hardware [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1104.496191] env[62569]: DEBUG nova.virt.hardware [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1104.496352] env[62569]: DEBUG nova.virt.hardware [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1104.496388] env[62569]: DEBUG nova.virt.hardware [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1104.496599] env[62569]: DEBUG nova.virt.hardware [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1104.496762] env[62569]: DEBUG nova.virt.hardware [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1104.496932] env[62569]: DEBUG nova.virt.hardware [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1104.497109] env[62569]: DEBUG nova.virt.hardware [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1104.497287] env[62569]: DEBUG nova.virt.hardware [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1104.498146] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f75b4c6b-f058-4634-b582-16ae1aa386b9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.505789] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1da5254e-f5d1-4f40-af54-7170781ed386 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.584031] env[62569]: DEBUG nova.scheduler.client.report [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1104.627805] env[62569]: DEBUG nova.compute.manager [req-08b1a613-65ec-4e1f-84fe-33f7bbd73c68 req-997d6c1b-ee74-40b0-8988-ad0d392feb2d service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Received event network-vif-plugged-26f272f3-4864-483c-a76f-3bcf9efa4ab8 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1104.628042] env[62569]: DEBUG oslo_concurrency.lockutils [req-08b1a613-65ec-4e1f-84fe-33f7bbd73c68 req-997d6c1b-ee74-40b0-8988-ad0d392feb2d service nova] Acquiring lock "27b14750-e349-47af-affe-02d10050aad1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.628254] env[62569]: DEBUG oslo_concurrency.lockutils [req-08b1a613-65ec-4e1f-84fe-33f7bbd73c68 req-997d6c1b-ee74-40b0-8988-ad0d392feb2d service nova] Lock "27b14750-e349-47af-affe-02d10050aad1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.628491] env[62569]: DEBUG oslo_concurrency.lockutils [req-08b1a613-65ec-4e1f-84fe-33f7bbd73c68 req-997d6c1b-ee74-40b0-8988-ad0d392feb2d service nova] Lock "27b14750-e349-47af-affe-02d10050aad1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.628609] env[62569]: DEBUG nova.compute.manager [req-08b1a613-65ec-4e1f-84fe-33f7bbd73c68 req-997d6c1b-ee74-40b0-8988-ad0d392feb2d service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] No waiting events found dispatching network-vif-plugged-26f272f3-4864-483c-a76f-3bcf9efa4ab8 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1104.628752] env[62569]: WARNING nova.compute.manager [req-08b1a613-65ec-4e1f-84fe-33f7bbd73c68 req-997d6c1b-ee74-40b0-8988-ad0d392feb2d service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Received unexpected event network-vif-plugged-26f272f3-4864-483c-a76f-3bcf9efa4ab8 for instance with vm_state building and task_state spawning. [ 1104.705059] env[62569]: DEBUG nova.network.neutron [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Successfully updated port: 26f272f3-4864-483c-a76f-3bcf9efa4ab8 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1105.151085] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-0db13343-9297-4770-894a-3a071a74d1ca tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Volume attach. Driver type: vmdk {{(pid=62569) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1105.151361] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-0db13343-9297-4770-894a-3a071a74d1ca tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269510', 'volume_id': 'b29ad30c-88ee-4fac-a8b0-ecc6276571dd', 'name': 'volume-b29ad30c-88ee-4fac-a8b0-ecc6276571dd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cf077eb8-8860-46f0-ae84-eed58c29873b', 'attached_at': '', 'detached_at': '', 'volume_id': 'b29ad30c-88ee-4fac-a8b0-ecc6276571dd', 'serial': 'b29ad30c-88ee-4fac-a8b0-ecc6276571dd'} {{(pid=62569) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1105.152279] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d19a800c-4ae2-4622-9767-ca681590bfe1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.168788] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d078626-4ecc-4543-9299-8a7b45d626ec {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.193877] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-0db13343-9297-4770-894a-3a071a74d1ca tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] volume-b29ad30c-88ee-4fac-a8b0-ecc6276571dd/volume-b29ad30c-88ee-4fac-a8b0-ecc6276571dd.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1105.194191] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b276789-f71f-4d36-aa16-4516a8f72655 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.207248] env[62569]: DEBUG oslo_concurrency.lockutils [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.207443] env[62569]: DEBUG oslo_concurrency.lockutils [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.207522] env[62569]: DEBUG nova.network.neutron [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1105.213627] env[62569]: DEBUG oslo_vmware.api [None req-0db13343-9297-4770-894a-3a071a74d1ca tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 1105.213627] env[62569]: value = "task-1250723" [ 1105.213627] env[62569]: _type = "Task" [ 1105.213627] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.222605] env[62569]: DEBUG oslo_vmware.api [None req-0db13343-9297-4770-894a-3a071a74d1ca tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250723, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.594943] env[62569]: DEBUG oslo_concurrency.lockutils [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.648s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.724793] env[62569]: DEBUG oslo_vmware.api [None req-0db13343-9297-4770-894a-3a071a74d1ca tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250723, 'name': ReconfigVM_Task, 'duration_secs': 0.295921} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.725112] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-0db13343-9297-4770-894a-3a071a74d1ca tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Reconfigured VM instance instance-0000006d to attach disk [datastore1] volume-b29ad30c-88ee-4fac-a8b0-ecc6276571dd/volume-b29ad30c-88ee-4fac-a8b0-ecc6276571dd.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1105.729660] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37b16bd0-62f8-4adb-80a1-da2785d2954d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.740032] env[62569]: DEBUG nova.network.neutron [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1105.747461] env[62569]: DEBUG oslo_vmware.api [None req-0db13343-9297-4770-894a-3a071a74d1ca tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 1105.747461] env[62569]: value = "task-1250724" [ 1105.747461] env[62569]: _type = "Task" [ 1105.747461] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.755310] env[62569]: DEBUG oslo_vmware.api [None req-0db13343-9297-4770-894a-3a071a74d1ca tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250724, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.866506] env[62569]: DEBUG nova.network.neutron [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Updating instance_info_cache with network_info: [{"id": "26f272f3-4864-483c-a76f-3bcf9efa4ab8", "address": "fa:16:3e:c8:1f:5b", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26f272f3-48", "ovs_interfaceid": "26f272f3-4864-483c-a76f-3bcf9efa4ab8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.257724] env[62569]: DEBUG oslo_vmware.api [None req-0db13343-9297-4770-894a-3a071a74d1ca tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250724, 'name': ReconfigVM_Task, 'duration_secs': 0.138609} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.258056] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-0db13343-9297-4770-894a-3a071a74d1ca tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269510', 'volume_id': 'b29ad30c-88ee-4fac-a8b0-ecc6276571dd', 'name': 'volume-b29ad30c-88ee-4fac-a8b0-ecc6276571dd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cf077eb8-8860-46f0-ae84-eed58c29873b', 'attached_at': '', 'detached_at': '', 'volume_id': 'b29ad30c-88ee-4fac-a8b0-ecc6276571dd', 'serial': 'b29ad30c-88ee-4fac-a8b0-ecc6276571dd'} {{(pid=62569) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1106.368975] env[62569]: DEBUG oslo_concurrency.lockutils [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Releasing lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.369336] env[62569]: DEBUG nova.compute.manager [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Instance network_info: |[{"id": "26f272f3-4864-483c-a76f-3bcf9efa4ab8", "address": "fa:16:3e:c8:1f:5b", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26f272f3-48", "ovs_interfaceid": "26f272f3-4864-483c-a76f-3bcf9efa4ab8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1106.370054] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c8:1f:5b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b356db78-99c7-4464-822c-fc7e193f7878', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '26f272f3-4864-483c-a76f-3bcf9efa4ab8', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1106.377255] env[62569]: DEBUG oslo.service.loopingcall [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1106.377482] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27b14750-e349-47af-affe-02d10050aad1] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1106.377704] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ed0c042c-3e9f-4da0-8421-981de9bf2be5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.398067] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1106.398067] env[62569]: value = "task-1250725" [ 1106.398067] env[62569]: _type = "Task" [ 1106.398067] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.406271] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250725, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.656298] env[62569]: DEBUG nova.compute.manager [req-54535348-f277-4aa2-ae19-9dc35ff54983 req-b5af7161-ebae-4491-ad1a-d00ed25d64b9 service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Received event network-changed-26f272f3-4864-483c-a76f-3bcf9efa4ab8 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1106.656535] env[62569]: DEBUG nova.compute.manager [req-54535348-f277-4aa2-ae19-9dc35ff54983 req-b5af7161-ebae-4491-ad1a-d00ed25d64b9 service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Refreshing instance network info cache due to event network-changed-26f272f3-4864-483c-a76f-3bcf9efa4ab8. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1106.656722] env[62569]: DEBUG oslo_concurrency.lockutils [req-54535348-f277-4aa2-ae19-9dc35ff54983 req-b5af7161-ebae-4491-ad1a-d00ed25d64b9 service nova] Acquiring lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1106.656877] env[62569]: DEBUG oslo_concurrency.lockutils [req-54535348-f277-4aa2-ae19-9dc35ff54983 req-b5af7161-ebae-4491-ad1a-d00ed25d64b9 service nova] Acquired lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.657084] env[62569]: DEBUG nova.network.neutron [req-54535348-f277-4aa2-ae19-9dc35ff54983 req-b5af7161-ebae-4491-ad1a-d00ed25d64b9 service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Refreshing network info cache for port 26f272f3-4864-483c-a76f-3bcf9efa4ab8 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1106.907697] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250725, 'name': CreateVM_Task, 'duration_secs': 0.293614} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.907832] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27b14750-e349-47af-affe-02d10050aad1] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1106.908508] env[62569]: DEBUG oslo_concurrency.lockutils [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1106.908684] env[62569]: DEBUG oslo_concurrency.lockutils [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.909037] env[62569]: DEBUG oslo_concurrency.lockutils [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1106.909291] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1a6d84f-ce31-4e5f-9079-20bf54ccadbd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.913633] env[62569]: DEBUG oslo_vmware.api [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1106.913633] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52fc3a0b-ea9f-0a28-282c-f9322a55ab1e" [ 1106.913633] env[62569]: _type = "Task" [ 1106.913633] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.921021] env[62569]: DEBUG oslo_vmware.api [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52fc3a0b-ea9f-0a28-282c-f9322a55ab1e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.136044] env[62569]: INFO nova.compute.manager [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Swapping old allocation on dict_keys(['fa06556a-5785-4014-b8bd-bc240a0cf716']) held by migration a3c050a1-397b-420f-b575-14e1ae1b7da5 for instance [ 1107.158840] env[62569]: DEBUG nova.scheduler.client.report [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Overwriting current allocation {'allocations': {'fa06556a-5785-4014-b8bd-bc240a0cf716': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 158}}, 'project_id': '7f988acdb9cd4ca28ca8916c2ede2db5', 'user_id': 'ccb9b3f4e38e4e81819d22e781661015', 'consumer_generation': 1} on consumer 96167629-8777-4a1c-b564-c70d1bb59b90 {{(pid=62569) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 1107.243027] env[62569]: DEBUG oslo_concurrency.lockutils [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "refresh_cache-96167629-8777-4a1c-b564-c70d1bb59b90" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.243248] env[62569]: DEBUG oslo_concurrency.lockutils [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "refresh_cache-96167629-8777-4a1c-b564-c70d1bb59b90" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.243432] env[62569]: DEBUG nova.network.neutron [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1107.292009] env[62569]: DEBUG nova.objects.instance [None req-0db13343-9297-4770-894a-3a071a74d1ca tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lazy-loading 'flavor' on Instance uuid cf077eb8-8860-46f0-ae84-eed58c29873b {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1107.364425] env[62569]: DEBUG nova.network.neutron [req-54535348-f277-4aa2-ae19-9dc35ff54983 req-b5af7161-ebae-4491-ad1a-d00ed25d64b9 service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Updated VIF entry in instance network info cache for port 26f272f3-4864-483c-a76f-3bcf9efa4ab8. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1107.364843] env[62569]: DEBUG nova.network.neutron [req-54535348-f277-4aa2-ae19-9dc35ff54983 req-b5af7161-ebae-4491-ad1a-d00ed25d64b9 service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Updating instance_info_cache with network_info: [{"id": "26f272f3-4864-483c-a76f-3bcf9efa4ab8", "address": "fa:16:3e:c8:1f:5b", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26f272f3-48", "ovs_interfaceid": "26f272f3-4864-483c-a76f-3bcf9efa4ab8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.423731] env[62569]: DEBUG oslo_vmware.api [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52fc3a0b-ea9f-0a28-282c-f9322a55ab1e, 'name': SearchDatastore_Task, 'duration_secs': 0.009955} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.424112] env[62569]: DEBUG oslo_concurrency.lockutils [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.424371] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1107.424620] env[62569]: DEBUG oslo_concurrency.lockutils [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.424812] env[62569]: DEBUG oslo_concurrency.lockutils [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.425104] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1107.425314] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a8020c28-b8cb-4a02-a222-2d4cdba45565 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.433393] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1107.433627] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1107.434346] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7486fe5c-1fa4-4638-a0e6-10d9a12f0975 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.439192] env[62569]: DEBUG oslo_vmware.api [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1107.439192] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]521f4f2e-4b26-ef3d-c168-952fc52c1fc4" [ 1107.439192] env[62569]: _type = "Task" [ 1107.439192] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.446939] env[62569]: DEBUG oslo_vmware.api [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]521f4f2e-4b26-ef3d-c168-952fc52c1fc4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.797290] env[62569]: DEBUG oslo_concurrency.lockutils [None req-0db13343-9297-4770-894a-3a071a74d1ca tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cf077eb8-8860-46f0-ae84-eed58c29873b" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.274s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.867221] env[62569]: DEBUG oslo_concurrency.lockutils [req-54535348-f277-4aa2-ae19-9dc35ff54983 req-b5af7161-ebae-4491-ad1a-d00ed25d64b9 service nova] Releasing lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.950430] env[62569]: DEBUG oslo_vmware.api [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]521f4f2e-4b26-ef3d-c168-952fc52c1fc4, 'name': SearchDatastore_Task, 'duration_secs': 0.007924} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.951392] env[62569]: DEBUG nova.network.neutron [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Updating instance_info_cache with network_info: [{"id": "d5f53bf1-303b-435e-941a-47c7ab293484", "address": "fa:16:3e:e8:8d:ca", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5f53bf1-30", "ovs_interfaceid": "d5f53bf1-303b-435e-941a-47c7ab293484", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.953414] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffd6a644-03c3-489e-b960-7e7dd18c4365 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.958743] env[62569]: DEBUG oslo_vmware.api [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1107.958743] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]525fb0ae-9af0-9d36-4299-34c2fab018bc" [ 1107.958743] env[62569]: _type = "Task" [ 1107.958743] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.962526] env[62569]: DEBUG oslo_concurrency.lockutils [None req-941e550b-5461-4c22-9bfa-c1a654b10300 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "cf077eb8-8860-46f0-ae84-eed58c29873b" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.962784] env[62569]: DEBUG oslo_concurrency.lockutils [None req-941e550b-5461-4c22-9bfa-c1a654b10300 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cf077eb8-8860-46f0-ae84-eed58c29873b" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.970128] env[62569]: DEBUG oslo_vmware.api [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]525fb0ae-9af0-9d36-4299-34c2fab018bc, 'name': SearchDatastore_Task, 'duration_secs': 0.008486} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.970407] env[62569]: DEBUG oslo_concurrency.lockutils [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.970706] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 27b14750-e349-47af-affe-02d10050aad1/27b14750-e349-47af-affe-02d10050aad1.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1107.970982] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c7dd9105-97ea-4ca1-86e8-079c70daad23 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.977831] env[62569]: DEBUG oslo_vmware.api [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1107.977831] env[62569]: value = "task-1250726" [ 1107.977831] env[62569]: _type = "Task" [ 1107.977831] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.985417] env[62569]: DEBUG oslo_vmware.api [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250726, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.454764] env[62569]: DEBUG oslo_concurrency.lockutils [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "refresh_cache-96167629-8777-4a1c-b564-c70d1bb59b90" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1108.455286] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1108.455588] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9bd97e80-46c2-4430-b872-ba2307ba18a5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.462554] env[62569]: DEBUG oslo_vmware.api [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1108.462554] env[62569]: value = "task-1250727" [ 1108.462554] env[62569]: _type = "Task" [ 1108.462554] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.465905] env[62569]: INFO nova.compute.manager [None req-941e550b-5461-4c22-9bfa-c1a654b10300 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Detaching volume b29ad30c-88ee-4fac-a8b0-ecc6276571dd [ 1108.472420] env[62569]: DEBUG oslo_vmware.api [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250727, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.486628] env[62569]: DEBUG oslo_vmware.api [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250726, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.412684} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.486890] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 27b14750-e349-47af-affe-02d10050aad1/27b14750-e349-47af-affe-02d10050aad1.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1108.487125] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1108.487393] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-155c0903-1399-45f1-9082-21d9eae2af89 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.494247] env[62569]: DEBUG oslo_vmware.api [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1108.494247] env[62569]: value = "task-1250728" [ 1108.494247] env[62569]: _type = "Task" [ 1108.494247] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.498791] env[62569]: INFO nova.virt.block_device [None req-941e550b-5461-4c22-9bfa-c1a654b10300 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Attempting to driver detach volume b29ad30c-88ee-4fac-a8b0-ecc6276571dd from mountpoint /dev/sdb [ 1108.499028] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-941e550b-5461-4c22-9bfa-c1a654b10300 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Volume detach. Driver type: vmdk {{(pid=62569) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1108.499226] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-941e550b-5461-4c22-9bfa-c1a654b10300 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269510', 'volume_id': 'b29ad30c-88ee-4fac-a8b0-ecc6276571dd', 'name': 'volume-b29ad30c-88ee-4fac-a8b0-ecc6276571dd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cf077eb8-8860-46f0-ae84-eed58c29873b', 'attached_at': '', 'detached_at': '', 'volume_id': 'b29ad30c-88ee-4fac-a8b0-ecc6276571dd', 'serial': 'b29ad30c-88ee-4fac-a8b0-ecc6276571dd'} {{(pid=62569) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1108.499959] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70df9081-7a46-4ab1-b9dc-0c0ab9f25e50 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.504763] env[62569]: DEBUG oslo_vmware.api [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250728, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.523767] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01a6ac48-668a-4074-8810-2bcc2897b9ad {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.530273] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-062eaf4d-c584-49ee-a5cf-6cabe04faef5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.549877] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b6f25ea-0bb5-4569-8b6e-e89eabfdcd46 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.564619] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-941e550b-5461-4c22-9bfa-c1a654b10300 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] The volume has not been displaced from its original location: [datastore1] volume-b29ad30c-88ee-4fac-a8b0-ecc6276571dd/volume-b29ad30c-88ee-4fac-a8b0-ecc6276571dd.vmdk. No consolidation needed. {{(pid=62569) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1108.569965] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-941e550b-5461-4c22-9bfa-c1a654b10300 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Reconfiguring VM instance instance-0000006d to detach disk 2001 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1108.570298] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-99fc6ab5-e7c4-4dc4-842f-1cac64338cbc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.588193] env[62569]: DEBUG oslo_vmware.api [None req-941e550b-5461-4c22-9bfa-c1a654b10300 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 1108.588193] env[62569]: value = "task-1250729" [ 1108.588193] env[62569]: _type = "Task" [ 1108.588193] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.596493] env[62569]: DEBUG oslo_vmware.api [None req-941e550b-5461-4c22-9bfa-c1a654b10300 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250729, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.972736] env[62569]: DEBUG oslo_vmware.api [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250727, 'name': PowerOffVM_Task, 'duration_secs': 0.207798} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.973095] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1108.973704] env[62569]: DEBUG nova.virt.hardware [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1108.973933] env[62569]: DEBUG nova.virt.hardware [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1108.974132] env[62569]: DEBUG nova.virt.hardware [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1108.974328] env[62569]: DEBUG nova.virt.hardware [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1108.974481] env[62569]: DEBUG nova.virt.hardware [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1108.974639] env[62569]: DEBUG nova.virt.hardware [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1108.974919] env[62569]: DEBUG nova.virt.hardware [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1108.975016] env[62569]: DEBUG nova.virt.hardware [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1108.975226] env[62569]: DEBUG nova.virt.hardware [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1108.975408] env[62569]: DEBUG nova.virt.hardware [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1108.975594] env[62569]: DEBUG nova.virt.hardware [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1108.980454] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67361189-c5b7-4ce0-89a0-6b8b071978e5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.994765] env[62569]: DEBUG oslo_vmware.api [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1108.994765] env[62569]: value = "task-1250730" [ 1108.994765] env[62569]: _type = "Task" [ 1108.994765] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.004338] env[62569]: DEBUG oslo_vmware.api [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250728, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.374316} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.007239] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1109.007521] env[62569]: DEBUG oslo_vmware.api [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250730, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.008197] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce36d1ea-6031-491f-a493-427b60246cbf {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.030034] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] 27b14750-e349-47af-affe-02d10050aad1/27b14750-e349-47af-affe-02d10050aad1.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1109.030341] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-869f60c8-aa62-4998-b80c-804e177c262b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.048753] env[62569]: DEBUG oslo_vmware.api [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1109.048753] env[62569]: value = "task-1250731" [ 1109.048753] env[62569]: _type = "Task" [ 1109.048753] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.055698] env[62569]: DEBUG oslo_vmware.api [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250731, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.097230] env[62569]: DEBUG oslo_vmware.api [None req-941e550b-5461-4c22-9bfa-c1a654b10300 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250729, 'name': ReconfigVM_Task, 'duration_secs': 0.299962} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.097507] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-941e550b-5461-4c22-9bfa-c1a654b10300 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Reconfigured VM instance instance-0000006d to detach disk 2001 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1109.102070] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-085279c6-0afd-4248-8ea0-adc7ad38bdf5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.116801] env[62569]: DEBUG oslo_vmware.api [None req-941e550b-5461-4c22-9bfa-c1a654b10300 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 1109.116801] env[62569]: value = "task-1250732" [ 1109.116801] env[62569]: _type = "Task" [ 1109.116801] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.127265] env[62569]: DEBUG oslo_vmware.api [None req-941e550b-5461-4c22-9bfa-c1a654b10300 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250732, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.506965] env[62569]: DEBUG oslo_vmware.api [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250730, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.558342] env[62569]: DEBUG oslo_vmware.api [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250731, 'name': ReconfigVM_Task, 'duration_secs': 0.260113} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.558573] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Reconfigured VM instance instance-0000006f to attach disk [datastore1] 27b14750-e349-47af-affe-02d10050aad1/27b14750-e349-47af-affe-02d10050aad1.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1109.559206] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ed178763-44aa-42ad-9e03-8763f500962c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.565190] env[62569]: DEBUG oslo_vmware.api [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1109.565190] env[62569]: value = "task-1250733" [ 1109.565190] env[62569]: _type = "Task" [ 1109.565190] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.574845] env[62569]: DEBUG oslo_vmware.api [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250733, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.625619] env[62569]: DEBUG oslo_vmware.api [None req-941e550b-5461-4c22-9bfa-c1a654b10300 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250732, 'name': ReconfigVM_Task, 'duration_secs': 0.13673} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.626748] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-941e550b-5461-4c22-9bfa-c1a654b10300 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-269510', 'volume_id': 'b29ad30c-88ee-4fac-a8b0-ecc6276571dd', 'name': 'volume-b29ad30c-88ee-4fac-a8b0-ecc6276571dd', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cf077eb8-8860-46f0-ae84-eed58c29873b', 'attached_at': '', 'detached_at': '', 'volume_id': 'b29ad30c-88ee-4fac-a8b0-ecc6276571dd', 'serial': 'b29ad30c-88ee-4fac-a8b0-ecc6276571dd'} {{(pid=62569) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1110.008075] env[62569]: DEBUG oslo_vmware.api [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250730, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.073744] env[62569]: DEBUG oslo_vmware.api [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250733, 'name': Rename_Task, 'duration_secs': 0.143478} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.074049] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1110.074298] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-35131f55-386d-4534-b0e9-3e0cdc34c970 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.079882] env[62569]: DEBUG oslo_vmware.api [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1110.079882] env[62569]: value = "task-1250734" [ 1110.079882] env[62569]: _type = "Task" [ 1110.079882] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.086832] env[62569]: DEBUG oslo_vmware.api [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250734, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.165923] env[62569]: DEBUG nova.objects.instance [None req-941e550b-5461-4c22-9bfa-c1a654b10300 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lazy-loading 'flavor' on Instance uuid cf077eb8-8860-46f0-ae84-eed58c29873b {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1110.508726] env[62569]: DEBUG oslo_vmware.api [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250730, 'name': ReconfigVM_Task, 'duration_secs': 1.152694} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.509695] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fea1793-e9e5-4094-8f67-ba49bc3fb38a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.527368] env[62569]: DEBUG nova.virt.hardware [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1110.527623] env[62569]: DEBUG nova.virt.hardware [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1110.527795] env[62569]: DEBUG nova.virt.hardware [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1110.527989] env[62569]: DEBUG nova.virt.hardware [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1110.528161] env[62569]: DEBUG nova.virt.hardware [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1110.528317] env[62569]: DEBUG nova.virt.hardware [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1110.528530] env[62569]: DEBUG nova.virt.hardware [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1110.528700] env[62569]: DEBUG nova.virt.hardware [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1110.528868] env[62569]: DEBUG nova.virt.hardware [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1110.529143] env[62569]: DEBUG nova.virt.hardware [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1110.529353] env[62569]: DEBUG nova.virt.hardware [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1110.530120] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e8b722f-08ae-4fb4-833b-347df876c472 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.535362] env[62569]: DEBUG oslo_vmware.api [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1110.535362] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529de8a1-c5c0-b6d4-dfe5-6c76e8200c60" [ 1110.535362] env[62569]: _type = "Task" [ 1110.535362] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.543387] env[62569]: DEBUG oslo_vmware.api [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529de8a1-c5c0-b6d4-dfe5-6c76e8200c60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.588812] env[62569]: DEBUG oslo_vmware.api [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250734, 'name': PowerOnVM_Task, 'duration_secs': 0.443185} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.589086] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1110.589293] env[62569]: INFO nova.compute.manager [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Took 6.12 seconds to spawn the instance on the hypervisor. [ 1110.589479] env[62569]: DEBUG nova.compute.manager [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1110.590236] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-623b75d1-8be4-4cf6-aff0-785bb4274b2a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.045260] env[62569]: DEBUG oslo_vmware.api [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529de8a1-c5c0-b6d4-dfe5-6c76e8200c60, 'name': SearchDatastore_Task, 'duration_secs': 0.00718} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.050510] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1111.050787] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1eb70f9f-bf1f-4913-a5c1-905ebf8a6c4a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.068415] env[62569]: DEBUG oslo_vmware.api [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1111.068415] env[62569]: value = "task-1250735" [ 1111.068415] env[62569]: _type = "Task" [ 1111.068415] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.077611] env[62569]: DEBUG oslo_vmware.api [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250735, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.106664] env[62569]: INFO nova.compute.manager [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Took 10.81 seconds to build instance. [ 1111.173511] env[62569]: DEBUG oslo_concurrency.lockutils [None req-941e550b-5461-4c22-9bfa-c1a654b10300 tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cf077eb8-8860-46f0-ae84-eed58c29873b" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.210s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.444749] env[62569]: DEBUG nova.compute.manager [req-ff4c4286-d46d-41c8-b708-9414da9d97ef req-44feb82e-b8bd-43cf-89a8-2d3170400ea0 service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Received event network-changed-26f272f3-4864-483c-a76f-3bcf9efa4ab8 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1111.444749] env[62569]: DEBUG nova.compute.manager [req-ff4c4286-d46d-41c8-b708-9414da9d97ef req-44feb82e-b8bd-43cf-89a8-2d3170400ea0 service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Refreshing instance network info cache due to event network-changed-26f272f3-4864-483c-a76f-3bcf9efa4ab8. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1111.444749] env[62569]: DEBUG oslo_concurrency.lockutils [req-ff4c4286-d46d-41c8-b708-9414da9d97ef req-44feb82e-b8bd-43cf-89a8-2d3170400ea0 service nova] Acquiring lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.444749] env[62569]: DEBUG oslo_concurrency.lockutils [req-ff4c4286-d46d-41c8-b708-9414da9d97ef req-44feb82e-b8bd-43cf-89a8-2d3170400ea0 service nova] Acquired lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.444902] env[62569]: DEBUG nova.network.neutron [req-ff4c4286-d46d-41c8-b708-9414da9d97ef req-44feb82e-b8bd-43cf-89a8-2d3170400ea0 service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Refreshing network info cache for port 26f272f3-4864-483c-a76f-3bcf9efa4ab8 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1111.578513] env[62569]: DEBUG oslo_vmware.api [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250735, 'name': ReconfigVM_Task, 'duration_secs': 0.161136} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.578793] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=62569) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1111.579568] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7e329a4-5de4-4640-914e-0a76de948c30 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.600932] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] 96167629-8777-4a1c-b564-c70d1bb59b90/96167629-8777-4a1c-b564-c70d1bb59b90.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1111.601197] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8268e671-133c-463b-9a5d-5bc3620d10d5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.613714] env[62569]: DEBUG oslo_concurrency.lockutils [None req-28043d06-82c5-4dac-a41f-38f4184be5fe tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "27b14750-e349-47af-affe-02d10050aad1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.326s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.618899] env[62569]: DEBUG oslo_vmware.api [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1111.618899] env[62569]: value = "task-1250736" [ 1111.618899] env[62569]: _type = "Task" [ 1111.618899] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.626239] env[62569]: DEBUG oslo_vmware.api [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250736, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.128853] env[62569]: DEBUG oslo_vmware.api [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250736, 'name': ReconfigVM_Task, 'duration_secs': 0.252943} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.128853] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Reconfigured VM instance instance-0000006c to attach disk [datastore2] 96167629-8777-4a1c-b564-c70d1bb59b90/96167629-8777-4a1c-b564-c70d1bb59b90.vmdk or device None with type thin {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1112.129294] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc2de5e8-367b-4ce6-8f3d-147f4a6a5510 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.148611] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38eb53db-9ee6-45c6-b0f1-953ea6c90482 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.167527] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f28c2989-f6fc-4d1b-9186-2b0c25111ded {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.187184] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbde43c5-fb3a-44d9-bf81-b70aed66cd17 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.194062] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1112.194306] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6b14a7b9-fde5-4b76-9e7a-10a2900f11a5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.199822] env[62569]: DEBUG oslo_vmware.api [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1112.199822] env[62569]: value = "task-1250737" [ 1112.199822] env[62569]: _type = "Task" [ 1112.199822] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.206969] env[62569]: DEBUG oslo_vmware.api [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250737, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.222914] env[62569]: DEBUG nova.network.neutron [req-ff4c4286-d46d-41c8-b708-9414da9d97ef req-44feb82e-b8bd-43cf-89a8-2d3170400ea0 service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Updated VIF entry in instance network info cache for port 26f272f3-4864-483c-a76f-3bcf9efa4ab8. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1112.223290] env[62569]: DEBUG nova.network.neutron [req-ff4c4286-d46d-41c8-b708-9414da9d97ef req-44feb82e-b8bd-43cf-89a8-2d3170400ea0 service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Updating instance_info_cache with network_info: [{"id": "26f272f3-4864-483c-a76f-3bcf9efa4ab8", "address": "fa:16:3e:c8:1f:5b", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26f272f3-48", "ovs_interfaceid": "26f272f3-4864-483c-a76f-3bcf9efa4ab8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.267713] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "cf077eb8-8860-46f0-ae84-eed58c29873b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.267956] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cf077eb8-8860-46f0-ae84-eed58c29873b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.268189] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "cf077eb8-8860-46f0-ae84-eed58c29873b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.268411] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cf077eb8-8860-46f0-ae84-eed58c29873b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.268658] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cf077eb8-8860-46f0-ae84-eed58c29873b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.270916] env[62569]: INFO nova.compute.manager [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Terminating instance [ 1112.709888] env[62569]: DEBUG oslo_vmware.api [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250737, 'name': PowerOnVM_Task, 'duration_secs': 0.351583} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.710175] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1112.725890] env[62569]: DEBUG oslo_concurrency.lockutils [req-ff4c4286-d46d-41c8-b708-9414da9d97ef req-44feb82e-b8bd-43cf-89a8-2d3170400ea0 service nova] Releasing lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1112.774389] env[62569]: DEBUG nova.compute.manager [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1112.774587] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1112.775495] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-563b7b73-607b-4294-8505-c57a579d3c1a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.783524] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1112.783782] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f3375286-9215-40e7-b624-51e1a456f60b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.790736] env[62569]: DEBUG oslo_vmware.api [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 1112.790736] env[62569]: value = "task-1250738" [ 1112.790736] env[62569]: _type = "Task" [ 1112.790736] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.799465] env[62569]: DEBUG oslo_vmware.api [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250738, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.300548] env[62569]: DEBUG oslo_vmware.api [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250738, 'name': PowerOffVM_Task, 'duration_secs': 0.194478} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.300973] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1113.301038] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1113.301270] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b70fd588-57db-461c-978c-9f7ed41855ea {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.367943] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1113.368191] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1113.368389] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Deleting the datastore file [datastore1] cf077eb8-8860-46f0-ae84-eed58c29873b {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1113.368681] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6c9b50a6-69c4-4923-abe6-7f3646680416 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.374893] env[62569]: DEBUG oslo_vmware.api [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for the task: (returnval){ [ 1113.374893] env[62569]: value = "task-1250740" [ 1113.374893] env[62569]: _type = "Task" [ 1113.374893] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.382920] env[62569]: DEBUG oslo_vmware.api [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250740, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.722269] env[62569]: INFO nova.compute.manager [None req-62325eef-2922-44f3-931d-1b10dad29f0d tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Updating instance to original state: 'active' [ 1113.887798] env[62569]: DEBUG oslo_vmware.api [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Task: {'id': task-1250740, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.122742} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.888154] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1113.888393] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1113.888617] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1113.888862] env[62569]: INFO nova.compute.manager [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1113.889211] env[62569]: DEBUG oslo.service.loopingcall [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1113.889456] env[62569]: DEBUG nova.compute.manager [-] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1113.889564] env[62569]: DEBUG nova.network.neutron [-] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1114.154444] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1114.402268] env[62569]: DEBUG nova.compute.manager [req-2b2d1488-5459-4199-b30a-31465de0edb1 req-d8350053-69e0-41cc-b5e0-eaaae3289703 service nova] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Received event network-vif-deleted-57d52bb3-5c0a-4ea0-89b3-94d652923656 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1114.402520] env[62569]: INFO nova.compute.manager [req-2b2d1488-5459-4199-b30a-31465de0edb1 req-d8350053-69e0-41cc-b5e0-eaaae3289703 service nova] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Neutron deleted interface 57d52bb3-5c0a-4ea0-89b3-94d652923656; detaching it from the instance and deleting it from the info cache [ 1114.402656] env[62569]: DEBUG nova.network.neutron [req-2b2d1488-5459-4199-b30a-31465de0edb1 req-d8350053-69e0-41cc-b5e0-eaaae3289703 service nova] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.881544] env[62569]: DEBUG nova.network.neutron [-] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.904687] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8aad7cfc-96cc-4a43-8f1d-69b42bc26dfd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.914633] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fe8e868-b7d1-4ed6-9fb8-560f46534992 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.939539] env[62569]: DEBUG nova.compute.manager [req-2b2d1488-5459-4199-b30a-31465de0edb1 req-d8350053-69e0-41cc-b5e0-eaaae3289703 service nova] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Detach interface failed, port_id=57d52bb3-5c0a-4ea0-89b3-94d652923656, reason: Instance cf077eb8-8860-46f0-ae84-eed58c29873b could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1115.170751] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "96167629-8777-4a1c-b564-c70d1bb59b90" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.170751] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "96167629-8777-4a1c-b564-c70d1bb59b90" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.170751] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "96167629-8777-4a1c-b564-c70d1bb59b90-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.170751] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "96167629-8777-4a1c-b564-c70d1bb59b90-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.170751] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "96167629-8777-4a1c-b564-c70d1bb59b90-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.171670] env[62569]: INFO nova.compute.manager [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Terminating instance [ 1115.384528] env[62569]: INFO nova.compute.manager [-] [instance: cf077eb8-8860-46f0-ae84-eed58c29873b] Took 1.49 seconds to deallocate network for instance. [ 1115.676827] env[62569]: DEBUG nova.compute.manager [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1115.677244] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1115.679516] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57e0179b-0723-4e7a-9e33-405dd5c3cda3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.689595] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1115.689931] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb769949-15ec-42a1-be7f-953226199ab5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.697338] env[62569]: DEBUG oslo_vmware.api [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1115.697338] env[62569]: value = "task-1250741" [ 1115.697338] env[62569]: _type = "Task" [ 1115.697338] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.709682] env[62569]: DEBUG oslo_vmware.api [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250741, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.891345] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.891671] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.891936] env[62569]: DEBUG nova.objects.instance [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lazy-loading 'resources' on Instance uuid cf077eb8-8860-46f0-ae84-eed58c29873b {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1116.207588] env[62569]: DEBUG oslo_vmware.api [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250741, 'name': PowerOffVM_Task, 'duration_secs': 0.170308} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.207850] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1116.208042] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1116.208300] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7ca43798-cfc1-4d21-9e72-826a26f36958 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.268969] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1116.269223] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1116.269422] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Deleting the datastore file [datastore2] 96167629-8777-4a1c-b564-c70d1bb59b90 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1116.269703] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5b2ad2c1-6e10-4b1f-9b1e-5a6b86197e27 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.275837] env[62569]: DEBUG oslo_vmware.api [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1116.275837] env[62569]: value = "task-1250743" [ 1116.275837] env[62569]: _type = "Task" [ 1116.275837] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.283701] env[62569]: DEBUG oslo_vmware.api [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250743, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.447774] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a0dc37-186f-4e80-a480-dbe381d32777 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.454938] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65deffff-cfc9-4424-ae5e-c5e91653d8bb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.483447] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5353636e-a3e9-4eb3-9d17-3df08d2ce53c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.490072] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c85100d-a8d1-451d-a957-d05cc8979b23 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.502439] env[62569]: DEBUG nova.compute.provider_tree [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1116.655522] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1116.785659] env[62569]: DEBUG oslo_vmware.api [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250743, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136497} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.786014] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1116.786066] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1116.786268] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1116.786458] env[62569]: INFO nova.compute.manager [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1116.786699] env[62569]: DEBUG oslo.service.loopingcall [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1116.786897] env[62569]: DEBUG nova.compute.manager [-] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1116.786989] env[62569]: DEBUG nova.network.neutron [-] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1117.005128] env[62569]: DEBUG nova.scheduler.client.report [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1117.477192] env[62569]: DEBUG nova.compute.manager [req-8101ea46-fd1c-46ad-8f00-baf440ce3791 req-26d19cd3-bd52-48a7-ba72-358f47ad48f8 service nova] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Received event network-vif-deleted-d5f53bf1-303b-435e-941a-47c7ab293484 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1117.477426] env[62569]: INFO nova.compute.manager [req-8101ea46-fd1c-46ad-8f00-baf440ce3791 req-26d19cd3-bd52-48a7-ba72-358f47ad48f8 service nova] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Neutron deleted interface d5f53bf1-303b-435e-941a-47c7ab293484; detaching it from the instance and deleting it from the info cache [ 1117.477608] env[62569]: DEBUG nova.network.neutron [req-8101ea46-fd1c-46ad-8f00-baf440ce3791 req-26d19cd3-bd52-48a7-ba72-358f47ad48f8 service nova] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1117.510623] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.619s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.532199] env[62569]: INFO nova.scheduler.client.report [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Deleted allocations for instance cf077eb8-8860-46f0-ae84-eed58c29873b [ 1117.953791] env[62569]: DEBUG nova.network.neutron [-] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1117.982613] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ec818b98-52a5-493a-8222-dc0d82cd48c3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.992570] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef2a23ad-b1b9-4f68-ab4d-8bfb17b546f3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.015929] env[62569]: DEBUG nova.compute.manager [req-8101ea46-fd1c-46ad-8f00-baf440ce3791 req-26d19cd3-bd52-48a7-ba72-358f47ad48f8 service nova] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Detach interface failed, port_id=d5f53bf1-303b-435e-941a-47c7ab293484, reason: Instance 96167629-8777-4a1c-b564-c70d1bb59b90 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1118.040959] env[62569]: DEBUG oslo_concurrency.lockutils [None req-32deacc7-b3de-4d2e-bc26-ece07b50faff tempest-AttachVolumeNegativeTest-1159559713 tempest-AttachVolumeNegativeTest-1159559713-project-member] Lock "cf077eb8-8860-46f0-ae84-eed58c29873b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.773s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.456377] env[62569]: INFO nova.compute.manager [-] [instance: 96167629-8777-4a1c-b564-c70d1bb59b90] Took 1.67 seconds to deallocate network for instance. [ 1118.963406] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.963693] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.963920] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.984917] env[62569]: INFO nova.scheduler.client.report [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Deleted allocations for instance 96167629-8777-4a1c-b564-c70d1bb59b90 [ 1119.494339] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bef7f112-5956-4982-8ec8-a71e75c579e7 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "96167629-8777-4a1c-b564-c70d1bb59b90" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.326s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.656027] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1119.656325] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1120.132054] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "4689e0e5-1566-4ec7-8060-ed89c5db6919" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.132054] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "4689e0e5-1566-4ec7-8060-ed89c5db6919" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.634184] env[62569]: DEBUG nova.compute.manager [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1120.655989] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1120.656301] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Starting heal instance info cache {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 1121.156295] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.156555] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.158097] env[62569]: INFO nova.compute.claims [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1122.208883] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10d69881-3c33-4308-b47d-4e79827b8375 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.216378] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-907bd1ee-2b83-44de-ba8e-1554aaa56b93 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.246526] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed26df58-5fe4-44ca-96b2-b2a79a725418 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.254249] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcd8cd04-bf7a-4c36-8616-874de985d4a3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.268634] env[62569]: DEBUG nova.compute.provider_tree [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1122.671850] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Didn't find any instances for network info cache update. {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10394}} [ 1122.672152] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.672299] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.672465] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.672677] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62569) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 1122.672759] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.771866] env[62569]: DEBUG nova.scheduler.client.report [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1123.176430] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.276601] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.120s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.277111] env[62569]: DEBUG nova.compute.manager [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1123.279839] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.104s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.280037] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.280230] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62569) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1123.281285] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b6ad977-8bff-4999-85bb-b9165ef0e5f0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.289621] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c35360d9-903c-48ac-befc-2fdb31016e75 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.303255] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0a69910-adb8-44af-8d66-bebd4f97f062 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.309474] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b24c2236-879f-469b-9b70-0f62002d67c8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.337781] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180871MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=62569) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1123.337940] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.338133] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.782998] env[62569]: DEBUG nova.compute.utils [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1123.784495] env[62569]: DEBUG nova.compute.manager [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Allocating IP information in the background. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1123.784667] env[62569]: DEBUG nova.network.neutron [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] allocate_for_instance() {{(pid=62569) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1123.828724] env[62569]: DEBUG nova.policy [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ccb9b3f4e38e4e81819d22e781661015', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7f988acdb9cd4ca28ca8916c2ede2db5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62569) authorize /opt/stack/nova/nova/policy.py:201}} [ 1124.092547] env[62569]: DEBUG nova.network.neutron [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Successfully created port: 44c168a4-d010-4c11-946f-b85089f76985 {{(pid=62569) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1124.287505] env[62569]: DEBUG nova.compute.manager [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1124.361307] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 27b14750-e349-47af-affe-02d10050aad1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1124.361425] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 4689e0e5-1566-4ec7-8060-ed89c5db6919 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1124.361519] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62569) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1124.361661] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62569) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1124.395454] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ddabc7c-0187-4c94-bead-db8960152484 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.402861] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73ffab10-2b6f-4a62-b110-c8491d6c177c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.431769] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d772def-ebca-489e-ab83-8cd9b7d8056e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.438996] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2ce2c1-3fb2-4cc9-ba8d-f4960bbbd5aa {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.451757] env[62569]: DEBUG nova.compute.provider_tree [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1124.954492] env[62569]: DEBUG nova.scheduler.client.report [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1125.296817] env[62569]: DEBUG nova.compute.manager [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1125.323535] env[62569]: DEBUG nova.virt.hardware [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1125.324159] env[62569]: DEBUG nova.virt.hardware [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1125.324159] env[62569]: DEBUG nova.virt.hardware [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1125.324332] env[62569]: DEBUG nova.virt.hardware [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1125.324332] env[62569]: DEBUG nova.virt.hardware [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1125.324720] env[62569]: DEBUG nova.virt.hardware [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1125.324720] env[62569]: DEBUG nova.virt.hardware [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1125.324850] env[62569]: DEBUG nova.virt.hardware [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1125.325123] env[62569]: DEBUG nova.virt.hardware [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1125.325256] env[62569]: DEBUG nova.virt.hardware [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1125.325447] env[62569]: DEBUG nova.virt.hardware [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1125.326342] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f21f1df-924d-4eaf-a898-0a4fbf330e1b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.334423] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1978e0fb-8df9-4b4e-a040-ac3d11721ae3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.459744] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62569) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1125.459965] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.122s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.495769] env[62569]: DEBUG nova.compute.manager [req-c9f7723a-f249-4a24-8f34-47a0f40bb1a4 req-acb67cae-1d8f-414a-9be8-56541fcd18a4 service nova] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Received event network-vif-plugged-44c168a4-d010-4c11-946f-b85089f76985 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1125.496094] env[62569]: DEBUG oslo_concurrency.lockutils [req-c9f7723a-f249-4a24-8f34-47a0f40bb1a4 req-acb67cae-1d8f-414a-9be8-56541fcd18a4 service nova] Acquiring lock "4689e0e5-1566-4ec7-8060-ed89c5db6919-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.496225] env[62569]: DEBUG oslo_concurrency.lockutils [req-c9f7723a-f249-4a24-8f34-47a0f40bb1a4 req-acb67cae-1d8f-414a-9be8-56541fcd18a4 service nova] Lock "4689e0e5-1566-4ec7-8060-ed89c5db6919-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.496405] env[62569]: DEBUG oslo_concurrency.lockutils [req-c9f7723a-f249-4a24-8f34-47a0f40bb1a4 req-acb67cae-1d8f-414a-9be8-56541fcd18a4 service nova] Lock "4689e0e5-1566-4ec7-8060-ed89c5db6919-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.496903] env[62569]: DEBUG nova.compute.manager [req-c9f7723a-f249-4a24-8f34-47a0f40bb1a4 req-acb67cae-1d8f-414a-9be8-56541fcd18a4 service nova] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] No waiting events found dispatching network-vif-plugged-44c168a4-d010-4c11-946f-b85089f76985 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1125.497043] env[62569]: WARNING nova.compute.manager [req-c9f7723a-f249-4a24-8f34-47a0f40bb1a4 req-acb67cae-1d8f-414a-9be8-56541fcd18a4 service nova] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Received unexpected event network-vif-plugged-44c168a4-d010-4c11-946f-b85089f76985 for instance with vm_state building and task_state spawning. [ 1125.559496] env[62569]: DEBUG nova.network.neutron [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Successfully updated port: 44c168a4-d010-4c11-946f-b85089f76985 {{(pid=62569) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1126.063332] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "refresh_cache-4689e0e5-1566-4ec7-8060-ed89c5db6919" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.063332] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "refresh_cache-4689e0e5-1566-4ec7-8060-ed89c5db6919" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.063332] env[62569]: DEBUG nova.network.neutron [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1126.595983] env[62569]: DEBUG nova.network.neutron [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1126.713374] env[62569]: DEBUG nova.network.neutron [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Updating instance_info_cache with network_info: [{"id": "44c168a4-d010-4c11-946f-b85089f76985", "address": "fa:16:3e:d7:2e:b2", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44c168a4-d0", "ovs_interfaceid": "44c168a4-d010-4c11-946f-b85089f76985", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1126.783882] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._sync_power_states {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1127.216508] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "refresh_cache-4689e0e5-1566-4ec7-8060-ed89c5db6919" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.216834] env[62569]: DEBUG nova.compute.manager [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Instance network_info: |[{"id": "44c168a4-d010-4c11-946f-b85089f76985", "address": "fa:16:3e:d7:2e:b2", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44c168a4-d0", "ovs_interfaceid": "44c168a4-d010-4c11-946f-b85089f76985", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1127.217324] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d7:2e:b2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73915082-a1b0-460b-b24d-97588fc9cb29', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '44c168a4-d010-4c11-946f-b85089f76985', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1127.224824] env[62569]: DEBUG oslo.service.loopingcall [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1127.225065] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1127.225325] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d3936e54-bdda-4b6b-9b42-a55995865ebb {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.244675] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1127.244675] env[62569]: value = "task-1250747" [ 1127.244675] env[62569]: _type = "Task" [ 1127.244675] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.251731] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250747, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.287031] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Getting list of instances from cluster (obj){ [ 1127.287031] env[62569]: value = "domain-c8" [ 1127.287031] env[62569]: _type = "ClusterComputeResource" [ 1127.287031] env[62569]: } {{(pid=62569) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1127.288058] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab46e75-7e66-499d-8186-2a2e8e698fbe {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.297799] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Got total of 1 instances {{(pid=62569) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1127.297961] env[62569]: WARNING nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] While synchronizing instance power states, found 2 instances in the database and 1 instances on the hypervisor. [ 1127.298122] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Triggering sync for uuid 27b14750-e349-47af-affe-02d10050aad1 {{(pid=62569) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10718}} [ 1127.298322] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Triggering sync for uuid 4689e0e5-1566-4ec7-8060-ed89c5db6919 {{(pid=62569) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10718}} [ 1127.298648] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "27b14750-e349-47af-affe-02d10050aad1" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.298876] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "27b14750-e349-47af-affe-02d10050aad1" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.299155] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "4689e0e5-1566-4ec7-8060-ed89c5db6919" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.299961] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5906ed10-ee77-4dad-aece-a2f627e3d731 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.522162] env[62569]: DEBUG nova.compute.manager [req-03e16435-56c7-41ec-bae5-9ff3b3a9a662 req-736010f2-6bc8-48d0-9e79-1b343551c437 service nova] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Received event network-changed-44c168a4-d010-4c11-946f-b85089f76985 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1127.522162] env[62569]: DEBUG nova.compute.manager [req-03e16435-56c7-41ec-bae5-9ff3b3a9a662 req-736010f2-6bc8-48d0-9e79-1b343551c437 service nova] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Refreshing instance network info cache due to event network-changed-44c168a4-d010-4c11-946f-b85089f76985. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1127.522162] env[62569]: DEBUG oslo_concurrency.lockutils [req-03e16435-56c7-41ec-bae5-9ff3b3a9a662 req-736010f2-6bc8-48d0-9e79-1b343551c437 service nova] Acquiring lock "refresh_cache-4689e0e5-1566-4ec7-8060-ed89c5db6919" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.522363] env[62569]: DEBUG oslo_concurrency.lockutils [req-03e16435-56c7-41ec-bae5-9ff3b3a9a662 req-736010f2-6bc8-48d0-9e79-1b343551c437 service nova] Acquired lock "refresh_cache-4689e0e5-1566-4ec7-8060-ed89c5db6919" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.522566] env[62569]: DEBUG nova.network.neutron [req-03e16435-56c7-41ec-bae5-9ff3b3a9a662 req-736010f2-6bc8-48d0-9e79-1b343551c437 service nova] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Refreshing network info cache for port 44c168a4-d010-4c11-946f-b85089f76985 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1127.758034] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250747, 'name': CreateVM_Task, 'duration_secs': 0.319959} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.758034] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1127.758034] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.758034] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.758034] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1127.758034] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d54da08a-3459-42de-ab16-53e56ed42066 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.762189] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1127.762189] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5249b67e-1909-a318-d576-460c58e10dbe" [ 1127.762189] env[62569]: _type = "Task" [ 1127.762189] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.769378] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5249b67e-1909-a318-d576-460c58e10dbe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.808843] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "27b14750-e349-47af-affe-02d10050aad1" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.510s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.274598] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5249b67e-1909-a318-d576-460c58e10dbe, 'name': SearchDatastore_Task, 'duration_secs': 0.010044} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.277179] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.277434] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1128.277669] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.277819] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.278000] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1128.278283] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-72e1d3c2-2226-4101-ba90-72d9978a8819 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.287214] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1128.287433] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1128.288149] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d1653f2-16c7-409d-a9ff-a3939e53deb0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.293031] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1128.293031] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52500a06-4f95-6a5b-7db2-7990c12b70ec" [ 1128.293031] env[62569]: _type = "Task" [ 1128.293031] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.300512] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52500a06-4f95-6a5b-7db2-7990c12b70ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.355568] env[62569]: DEBUG nova.network.neutron [req-03e16435-56c7-41ec-bae5-9ff3b3a9a662 req-736010f2-6bc8-48d0-9e79-1b343551c437 service nova] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Updated VIF entry in instance network info cache for port 44c168a4-d010-4c11-946f-b85089f76985. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1128.356652] env[62569]: DEBUG nova.network.neutron [req-03e16435-56c7-41ec-bae5-9ff3b3a9a662 req-736010f2-6bc8-48d0-9e79-1b343551c437 service nova] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Updating instance_info_cache with network_info: [{"id": "44c168a4-d010-4c11-946f-b85089f76985", "address": "fa:16:3e:d7:2e:b2", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44c168a4-d0", "ovs_interfaceid": "44c168a4-d010-4c11-946f-b85089f76985", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.806209] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52500a06-4f95-6a5b-7db2-7990c12b70ec, 'name': SearchDatastore_Task, 'duration_secs': 0.007912} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.807340] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8b8b295-3809-4bdb-99c3-1a7d5194f618 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.814295] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1128.814295] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b53378-2077-eb9b-9139-6e48a8c372f3" [ 1128.814295] env[62569]: _type = "Task" [ 1128.814295] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.822843] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b53378-2077-eb9b-9139-6e48a8c372f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.858895] env[62569]: DEBUG oslo_concurrency.lockutils [req-03e16435-56c7-41ec-bae5-9ff3b3a9a662 req-736010f2-6bc8-48d0-9e79-1b343551c437 service nova] Releasing lock "refresh_cache-4689e0e5-1566-4ec7-8060-ed89c5db6919" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.326749] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52b53378-2077-eb9b-9139-6e48a8c372f3, 'name': SearchDatastore_Task, 'duration_secs': 0.009113} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.326971] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.327401] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 4689e0e5-1566-4ec7-8060-ed89c5db6919/4689e0e5-1566-4ec7-8060-ed89c5db6919.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1129.327674] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f8d19cb3-5eca-49a9-8874-f8c6d5dccd09 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.335672] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1129.335672] env[62569]: value = "task-1250748" [ 1129.335672] env[62569]: _type = "Task" [ 1129.335672] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.343482] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250748, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.845812] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250748, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474301} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.845812] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 4689e0e5-1566-4ec7-8060-ed89c5db6919/4689e0e5-1566-4ec7-8060-ed89c5db6919.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1129.846247] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1129.846247] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fe467297-f68c-4e86-b9cb-da9f6fd76428 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.852759] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1129.852759] env[62569]: value = "task-1250749" [ 1129.852759] env[62569]: _type = "Task" [ 1129.852759] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.860555] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250749, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.362855] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250749, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068348} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.363141] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1130.363941] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4326744-f7e5-4cd6-9377-81a124061284 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.386311] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] 4689e0e5-1566-4ec7-8060-ed89c5db6919/4689e0e5-1566-4ec7-8060-ed89c5db6919.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1130.386551] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-84927a79-e175-4b99-b459-c896903a5736 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.405167] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1130.405167] env[62569]: value = "task-1250750" [ 1130.405167] env[62569]: _type = "Task" [ 1130.405167] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.413720] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250750, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.915107] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250750, 'name': ReconfigVM_Task, 'duration_secs': 0.29008} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.915476] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Reconfigured VM instance instance-00000070 to attach disk [datastore1] 4689e0e5-1566-4ec7-8060-ed89c5db6919/4689e0e5-1566-4ec7-8060-ed89c5db6919.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1130.916067] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f0cbb57d-e019-48aa-94fe-da0939044593 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.922548] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1130.922548] env[62569]: value = "task-1250751" [ 1130.922548] env[62569]: _type = "Task" [ 1130.922548] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.929992] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250751, 'name': Rename_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.432315] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250751, 'name': Rename_Task, 'duration_secs': 0.137581} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.432557] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1131.432811] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-315ce36a-ccdd-484c-89a4-d7b3ad74301f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.438808] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1131.438808] env[62569]: value = "task-1250752" [ 1131.438808] env[62569]: _type = "Task" [ 1131.438808] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.445862] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250752, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.949138] env[62569]: DEBUG oslo_vmware.api [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250752, 'name': PowerOnVM_Task, 'duration_secs': 0.453408} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.949514] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1131.949735] env[62569]: INFO nova.compute.manager [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Took 6.65 seconds to spawn the instance on the hypervisor. [ 1131.949969] env[62569]: DEBUG nova.compute.manager [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1131.950833] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b840d8d3-53fb-4aaf-85de-717caa29c766 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.468289] env[62569]: INFO nova.compute.manager [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Took 11.33 seconds to build instance. [ 1132.970219] env[62569]: DEBUG oslo_concurrency.lockutils [None req-bc78cc2a-ad30-4225-96c6-7c0f4c205d03 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "4689e0e5-1566-4ec7-8060-ed89c5db6919" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.839s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.970575] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "4689e0e5-1566-4ec7-8060-ed89c5db6919" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 5.671s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.970742] env[62569]: INFO nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] During sync_power_state the instance has a pending task (spawning). Skip. [ 1132.970926] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "4689e0e5-1566-4ec7-8060-ed89c5db6919" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.333574] env[62569]: DEBUG nova.compute.manager [req-5e041afb-eed9-4104-8dc3-7a432185dfff req-38208d3a-408a-43b1-85bb-2b768b07316a service nova] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Received event network-changed-44c168a4-d010-4c11-946f-b85089f76985 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1133.333716] env[62569]: DEBUG nova.compute.manager [req-5e041afb-eed9-4104-8dc3-7a432185dfff req-38208d3a-408a-43b1-85bb-2b768b07316a service nova] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Refreshing instance network info cache due to event network-changed-44c168a4-d010-4c11-946f-b85089f76985. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1133.333941] env[62569]: DEBUG oslo_concurrency.lockutils [req-5e041afb-eed9-4104-8dc3-7a432185dfff req-38208d3a-408a-43b1-85bb-2b768b07316a service nova] Acquiring lock "refresh_cache-4689e0e5-1566-4ec7-8060-ed89c5db6919" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.334178] env[62569]: DEBUG oslo_concurrency.lockutils [req-5e041afb-eed9-4104-8dc3-7a432185dfff req-38208d3a-408a-43b1-85bb-2b768b07316a service nova] Acquired lock "refresh_cache-4689e0e5-1566-4ec7-8060-ed89c5db6919" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.334341] env[62569]: DEBUG nova.network.neutron [req-5e041afb-eed9-4104-8dc3-7a432185dfff req-38208d3a-408a-43b1-85bb-2b768b07316a service nova] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Refreshing network info cache for port 44c168a4-d010-4c11-946f-b85089f76985 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1134.107210] env[62569]: DEBUG nova.network.neutron [req-5e041afb-eed9-4104-8dc3-7a432185dfff req-38208d3a-408a-43b1-85bb-2b768b07316a service nova] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Updated VIF entry in instance network info cache for port 44c168a4-d010-4c11-946f-b85089f76985. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1134.107603] env[62569]: DEBUG nova.network.neutron [req-5e041afb-eed9-4104-8dc3-7a432185dfff req-38208d3a-408a-43b1-85bb-2b768b07316a service nova] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Updating instance_info_cache with network_info: [{"id": "44c168a4-d010-4c11-946f-b85089f76985", "address": "fa:16:3e:d7:2e:b2", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44c168a4-d0", "ovs_interfaceid": "44c168a4-d010-4c11-946f-b85089f76985", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.610372] env[62569]: DEBUG oslo_concurrency.lockutils [req-5e041afb-eed9-4104-8dc3-7a432185dfff req-38208d3a-408a-43b1-85bb-2b768b07316a service nova] Releasing lock "refresh_cache-4689e0e5-1566-4ec7-8060-ed89c5db6919" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1149.159772] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "27b14750-e349-47af-affe-02d10050aad1" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1149.160241] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "27b14750-e349-47af-affe-02d10050aad1" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.160288] env[62569]: INFO nova.compute.manager [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Shelving [ 1150.168948] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1150.169298] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-004f511e-4e25-4a4e-8742-80f408b253f7 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.176491] env[62569]: DEBUG oslo_vmware.api [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1150.176491] env[62569]: value = "task-1250753" [ 1150.176491] env[62569]: _type = "Task" [ 1150.176491] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.184221] env[62569]: DEBUG oslo_vmware.api [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250753, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.686509] env[62569]: DEBUG oslo_vmware.api [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250753, 'name': PowerOffVM_Task, 'duration_secs': 0.243172} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.686775] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1150.687579] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcefe808-76ed-4944-8077-b592c227b49b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.705594] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce6b90b9-d25c-4b5b-941a-b1619d632b96 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.215928] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Creating Snapshot of the VM instance {{(pid=62569) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1151.216440] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-cc8ade10-fa7c-4bc9-b1ba-85971d268437 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.225190] env[62569]: DEBUG oslo_vmware.api [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1151.225190] env[62569]: value = "task-1250754" [ 1151.225190] env[62569]: _type = "Task" [ 1151.225190] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.233592] env[62569]: DEBUG oslo_vmware.api [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250754, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.735518] env[62569]: DEBUG oslo_vmware.api [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250754, 'name': CreateSnapshot_Task, 'duration_secs': 0.401186} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.735794] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Created Snapshot of the VM instance {{(pid=62569) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1151.736542] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75f3e25c-5a33-46db-b73f-4ed0afe4015d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.252997] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Creating linked-clone VM from snapshot {{(pid=62569) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1152.253378] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b22dacbf-19ce-4822-a640-d7ccc4f1d6cd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.261555] env[62569]: DEBUG oslo_vmware.api [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1152.261555] env[62569]: value = "task-1250755" [ 1152.261555] env[62569]: _type = "Task" [ 1152.261555] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.269218] env[62569]: DEBUG oslo_vmware.api [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250755, 'name': CloneVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.771843] env[62569]: DEBUG oslo_vmware.api [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250755, 'name': CloneVM_Task} progress is 94%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.271807] env[62569]: DEBUG oslo_vmware.api [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250755, 'name': CloneVM_Task, 'duration_secs': 0.946549} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.272123] env[62569]: INFO nova.virt.vmwareapi.vmops [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Created linked-clone VM from snapshot [ 1153.272829] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6075607f-8cc3-489d-b11c-cef8b9f5ab41 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.281104] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Uploading image 7d09e0bb-738b-4c70-8d49-3e9198c8c1f2 {{(pid=62569) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1153.300609] env[62569]: DEBUG oslo_vmware.rw_handles [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1153.300609] env[62569]: value = "vm-269514" [ 1153.300609] env[62569]: _type = "VirtualMachine" [ 1153.300609] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1153.300848] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-0737caf8-a81c-43b8-9304-92842a8e3819 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.306747] env[62569]: DEBUG oslo_vmware.rw_handles [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lease: (returnval){ [ 1153.306747] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52760add-b8c5-6192-c3a0-5334a38b66a1" [ 1153.306747] env[62569]: _type = "HttpNfcLease" [ 1153.306747] env[62569]: } obtained for exporting VM: (result){ [ 1153.306747] env[62569]: value = "vm-269514" [ 1153.306747] env[62569]: _type = "VirtualMachine" [ 1153.306747] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1153.307044] env[62569]: DEBUG oslo_vmware.api [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the lease: (returnval){ [ 1153.307044] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52760add-b8c5-6192-c3a0-5334a38b66a1" [ 1153.307044] env[62569]: _type = "HttpNfcLease" [ 1153.307044] env[62569]: } to be ready. {{(pid=62569) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1153.312691] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1153.312691] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52760add-b8c5-6192-c3a0-5334a38b66a1" [ 1153.312691] env[62569]: _type = "HttpNfcLease" [ 1153.312691] env[62569]: } is initializing. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1153.814879] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1153.814879] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52760add-b8c5-6192-c3a0-5334a38b66a1" [ 1153.814879] env[62569]: _type = "HttpNfcLease" [ 1153.814879] env[62569]: } is ready. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1153.815298] env[62569]: DEBUG oslo_vmware.rw_handles [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1153.815298] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52760add-b8c5-6192-c3a0-5334a38b66a1" [ 1153.815298] env[62569]: _type = "HttpNfcLease" [ 1153.815298] env[62569]: }. {{(pid=62569) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1153.815962] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f50244-226c-4dc9-ba95-68a863ed4d2e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.822867] env[62569]: DEBUG oslo_vmware.rw_handles [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525ab2ac-417c-256c-6acc-10fece5914e3/disk-0.vmdk from lease info. {{(pid=62569) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1153.823057] env[62569]: DEBUG oslo_vmware.rw_handles [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525ab2ac-417c-256c-6acc-10fece5914e3/disk-0.vmdk for reading. {{(pid=62569) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1153.910591] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b399a440-24bd-4d67-90dd-b2a004d5f61b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.914569] env[62569]: DEBUG oslo_vmware.rw_handles [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525ab2ac-417c-256c-6acc-10fece5914e3/disk-0.vmdk. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1161.915587] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf9c7ec4-306e-46f8-9d17-fd87d8960ab9 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.921627] env[62569]: DEBUG oslo_vmware.rw_handles [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525ab2ac-417c-256c-6acc-10fece5914e3/disk-0.vmdk is in state: ready. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1161.921797] env[62569]: ERROR oslo_vmware.rw_handles [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525ab2ac-417c-256c-6acc-10fece5914e3/disk-0.vmdk due to incomplete transfer. [ 1161.922016] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7ff842ce-8c77-41db-9191-065f48ae6536 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.929515] env[62569]: DEBUG oslo_vmware.rw_handles [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525ab2ac-417c-256c-6acc-10fece5914e3/disk-0.vmdk. {{(pid=62569) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1161.929716] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Uploaded image 7d09e0bb-738b-4c70-8d49-3e9198c8c1f2 to the Glance image server {{(pid=62569) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1161.931795] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Destroying the VM {{(pid=62569) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1161.932032] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8350aa3d-6307-4558-b8e1-891c2609cd4c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.937998] env[62569]: DEBUG oslo_vmware.api [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1161.937998] env[62569]: value = "task-1250757" [ 1161.937998] env[62569]: _type = "Task" [ 1161.937998] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.945153] env[62569]: DEBUG oslo_vmware.api [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250757, 'name': Destroy_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.447479] env[62569]: DEBUG oslo_vmware.api [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250757, 'name': Destroy_Task} progress is 33%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.948928] env[62569]: DEBUG oslo_vmware.api [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250757, 'name': Destroy_Task, 'duration_secs': 0.66171} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.949305] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Destroyed the VM [ 1162.949481] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Deleting Snapshot of the VM instance {{(pid=62569) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1162.949734] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1d2f1ea6-d547-430e-95b7-eb04dcbc3c45 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.955687] env[62569]: DEBUG oslo_vmware.api [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1162.955687] env[62569]: value = "task-1250758" [ 1162.955687] env[62569]: _type = "Task" [ 1162.955687] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.962867] env[62569]: DEBUG oslo_vmware.api [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250758, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.465434] env[62569]: DEBUG oslo_vmware.api [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250758, 'name': RemoveSnapshot_Task, 'duration_secs': 0.33861} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.465711] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Deleted Snapshot of the VM instance {{(pid=62569) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1163.466032] env[62569]: DEBUG nova.compute.manager [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1163.466822] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f04a14b-c86a-41fe-90b3-1c04a5e68478 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.978614] env[62569]: INFO nova.compute.manager [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Shelve offloading [ 1164.482749] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1164.483062] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-689d6ad5-0f35-4d10-9797-c16acce69552 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.491227] env[62569]: DEBUG oslo_vmware.api [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1164.491227] env[62569]: value = "task-1250759" [ 1164.491227] env[62569]: _type = "Task" [ 1164.491227] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.500815] env[62569]: DEBUG oslo_vmware.api [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250759, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.001518] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] VM already powered off {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1165.001807] env[62569]: DEBUG nova.compute.manager [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1165.002409] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-372eba6e-5664-43bc-9238-ea4e49101b8a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.007924] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1165.008108] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1165.008289] env[62569]: DEBUG nova.network.neutron [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1165.711859] env[62569]: DEBUG nova.network.neutron [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Updating instance_info_cache with network_info: [{"id": "26f272f3-4864-483c-a76f-3bcf9efa4ab8", "address": "fa:16:3e:c8:1f:5b", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26f272f3-48", "ovs_interfaceid": "26f272f3-4864-483c-a76f-3bcf9efa4ab8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.214953] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Releasing lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1166.418314] env[62569]: DEBUG nova.compute.manager [req-c75f50d2-b663-4524-b002-a999dd5d1df9 req-fa4e280f-54df-42e5-80fd-58467e6fac9c service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Received event network-vif-unplugged-26f272f3-4864-483c-a76f-3bcf9efa4ab8 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1166.418447] env[62569]: DEBUG oslo_concurrency.lockutils [req-c75f50d2-b663-4524-b002-a999dd5d1df9 req-fa4e280f-54df-42e5-80fd-58467e6fac9c service nova] Acquiring lock "27b14750-e349-47af-affe-02d10050aad1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.418657] env[62569]: DEBUG oslo_concurrency.lockutils [req-c75f50d2-b663-4524-b002-a999dd5d1df9 req-fa4e280f-54df-42e5-80fd-58467e6fac9c service nova] Lock "27b14750-e349-47af-affe-02d10050aad1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1166.418828] env[62569]: DEBUG oslo_concurrency.lockutils [req-c75f50d2-b663-4524-b002-a999dd5d1df9 req-fa4e280f-54df-42e5-80fd-58467e6fac9c service nova] Lock "27b14750-e349-47af-affe-02d10050aad1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1166.419010] env[62569]: DEBUG nova.compute.manager [req-c75f50d2-b663-4524-b002-a999dd5d1df9 req-fa4e280f-54df-42e5-80fd-58467e6fac9c service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] No waiting events found dispatching network-vif-unplugged-26f272f3-4864-483c-a76f-3bcf9efa4ab8 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1166.419200] env[62569]: WARNING nova.compute.manager [req-c75f50d2-b663-4524-b002-a999dd5d1df9 req-fa4e280f-54df-42e5-80fd-58467e6fac9c service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Received unexpected event network-vif-unplugged-26f272f3-4864-483c-a76f-3bcf9efa4ab8 for instance with vm_state shelved and task_state shelving_offloading. [ 1166.542428] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1166.543399] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eef51b69-550b-480a-a851-95dfd633e4ee {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.551049] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1166.551288] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-75421cef-50b0-492e-9186-0733832881aa {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.614711] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1166.614962] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1166.615179] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Deleting the datastore file [datastore1] 27b14750-e349-47af-affe-02d10050aad1 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1166.615449] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-545ffaaa-1871-469a-a473-d3553c0cf71a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.621661] env[62569]: DEBUG oslo_vmware.api [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1166.621661] env[62569]: value = "task-1250761" [ 1166.621661] env[62569]: _type = "Task" [ 1166.621661] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.628897] env[62569]: DEBUG oslo_vmware.api [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250761, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.131884] env[62569]: DEBUG oslo_vmware.api [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250761, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141353} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.132693] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1167.132693] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1167.132693] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1167.169489] env[62569]: INFO nova.scheduler.client.report [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Deleted allocations for instance 27b14750-e349-47af-affe-02d10050aad1 [ 1167.674228] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1167.674567] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1167.674893] env[62569]: DEBUG nova.objects.instance [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lazy-loading 'resources' on Instance uuid 27b14750-e349-47af-affe-02d10050aad1 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1168.178995] env[62569]: DEBUG nova.objects.instance [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lazy-loading 'numa_topology' on Instance uuid 27b14750-e349-47af-affe-02d10050aad1 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1168.199690] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b808eb47-d1b3-4ea2-b6c1-b1bb5fd60460 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "4689e0e5-1566-4ec7-8060-ed89c5db6919" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.199939] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b808eb47-d1b3-4ea2-b6c1-b1bb5fd60460 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "4689e0e5-1566-4ec7-8060-ed89c5db6919" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.200218] env[62569]: DEBUG nova.compute.manager [None req-b808eb47-d1b3-4ea2-b6c1-b1bb5fd60460 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1168.201182] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bee680f-41f3-474f-b4cb-e02346dadfef {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.208109] env[62569]: DEBUG nova.compute.manager [None req-b808eb47-d1b3-4ea2-b6c1-b1bb5fd60460 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62569) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1168.208673] env[62569]: DEBUG nova.objects.instance [None req-b808eb47-d1b3-4ea2-b6c1-b1bb5fd60460 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lazy-loading 'flavor' on Instance uuid 4689e0e5-1566-4ec7-8060-ed89c5db6919 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1168.443231] env[62569]: DEBUG nova.compute.manager [req-e1298b18-c8c2-4cc6-a21e-8e45fc99c036 req-60717680-17ae-43b6-92a5-19269951204e service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Received event network-changed-26f272f3-4864-483c-a76f-3bcf9efa4ab8 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1168.443450] env[62569]: DEBUG nova.compute.manager [req-e1298b18-c8c2-4cc6-a21e-8e45fc99c036 req-60717680-17ae-43b6-92a5-19269951204e service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Refreshing instance network info cache due to event network-changed-26f272f3-4864-483c-a76f-3bcf9efa4ab8. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1168.443672] env[62569]: DEBUG oslo_concurrency.lockutils [req-e1298b18-c8c2-4cc6-a21e-8e45fc99c036 req-60717680-17ae-43b6-92a5-19269951204e service nova] Acquiring lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1168.443841] env[62569]: DEBUG oslo_concurrency.lockutils [req-e1298b18-c8c2-4cc6-a21e-8e45fc99c036 req-60717680-17ae-43b6-92a5-19269951204e service nova] Acquired lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1168.443984] env[62569]: DEBUG nova.network.neutron [req-e1298b18-c8c2-4cc6-a21e-8e45fc99c036 req-60717680-17ae-43b6-92a5-19269951204e service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Refreshing network info cache for port 26f272f3-4864-483c-a76f-3bcf9efa4ab8 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1168.681362] env[62569]: DEBUG nova.objects.base [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Object Instance<27b14750-e349-47af-affe-02d10050aad1> lazy-loaded attributes: resources,numa_topology {{(pid=62569) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1168.719448] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7520d40-0e07-4ffe-ba78-347873c252c2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.727882] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8155cc2-6bc3-4024-96d6-ce05a93e89ce {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.756618] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-707de354-8e46-4219-899d-ee66b522dd10 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.763243] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e18697b2-ebf1-4fec-a61d-9373896b64af {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.775762] env[62569]: DEBUG nova.compute.provider_tree [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1169.125557] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "27b14750-e349-47af-affe-02d10050aad1" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.215145] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-b808eb47-d1b3-4ea2-b6c1-b1bb5fd60460 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1169.215478] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5417115b-18c8-49ef-8831-3db1fcad0712 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.223418] env[62569]: DEBUG oslo_vmware.api [None req-b808eb47-d1b3-4ea2-b6c1-b1bb5fd60460 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1169.223418] env[62569]: value = "task-1250762" [ 1169.223418] env[62569]: _type = "Task" [ 1169.223418] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.231609] env[62569]: DEBUG oslo_vmware.api [None req-b808eb47-d1b3-4ea2-b6c1-b1bb5fd60460 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250762, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.278048] env[62569]: DEBUG nova.scheduler.client.report [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1169.383190] env[62569]: DEBUG nova.network.neutron [req-e1298b18-c8c2-4cc6-a21e-8e45fc99c036 req-60717680-17ae-43b6-92a5-19269951204e service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Updated VIF entry in instance network info cache for port 26f272f3-4864-483c-a76f-3bcf9efa4ab8. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1169.383639] env[62569]: DEBUG nova.network.neutron [req-e1298b18-c8c2-4cc6-a21e-8e45fc99c036 req-60717680-17ae-43b6-92a5-19269951204e service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Updating instance_info_cache with network_info: [{"id": "26f272f3-4864-483c-a76f-3bcf9efa4ab8", "address": "fa:16:3e:c8:1f:5b", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": null, "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap26f272f3-48", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1169.734535] env[62569]: DEBUG oslo_vmware.api [None req-b808eb47-d1b3-4ea2-b6c1-b1bb5fd60460 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250762, 'name': PowerOffVM_Task, 'duration_secs': 0.20185} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.734905] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-b808eb47-d1b3-4ea2-b6c1-b1bb5fd60460 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1169.735041] env[62569]: DEBUG nova.compute.manager [None req-b808eb47-d1b3-4ea2-b6c1-b1bb5fd60460 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1169.735792] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67cf51ba-b3ba-409d-aaae-bd2d4325d952 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.786807] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.112s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.886063] env[62569]: DEBUG oslo_concurrency.lockutils [req-e1298b18-c8c2-4cc6-a21e-8e45fc99c036 req-60717680-17ae-43b6-92a5-19269951204e service nova] Releasing lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1170.246723] env[62569]: DEBUG oslo_concurrency.lockutils [None req-b808eb47-d1b3-4ea2-b6c1-b1bb5fd60460 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "4689e0e5-1566-4ec7-8060-ed89c5db6919" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.047s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.293300] env[62569]: DEBUG oslo_concurrency.lockutils [None req-4dd1aa97-8177-4dfb-9e7c-42c04e4ab778 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "27b14750-e349-47af-affe-02d10050aad1" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.133s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.293993] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "27b14750-e349-47af-affe-02d10050aad1" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.169s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.294200] env[62569]: INFO nova.compute.manager [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Unshelving [ 1170.602772] env[62569]: DEBUG nova.objects.instance [None req-63d1e0c6-b93e-4aa1-9912-8b422733ffe4 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lazy-loading 'flavor' on Instance uuid 4689e0e5-1566-4ec7-8060-ed89c5db6919 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1171.108178] env[62569]: DEBUG oslo_concurrency.lockutils [None req-63d1e0c6-b93e-4aa1-9912-8b422733ffe4 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "refresh_cache-4689e0e5-1566-4ec7-8060-ed89c5db6919" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1171.108550] env[62569]: DEBUG oslo_concurrency.lockutils [None req-63d1e0c6-b93e-4aa1-9912-8b422733ffe4 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "refresh_cache-4689e0e5-1566-4ec7-8060-ed89c5db6919" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1171.108550] env[62569]: DEBUG nova.network.neutron [None req-63d1e0c6-b93e-4aa1-9912-8b422733ffe4 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1171.108708] env[62569]: DEBUG nova.objects.instance [None req-63d1e0c6-b93e-4aa1-9912-8b422733ffe4 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lazy-loading 'info_cache' on Instance uuid 4689e0e5-1566-4ec7-8060-ed89c5db6919 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1171.321102] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.321388] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1171.321606] env[62569]: DEBUG nova.objects.instance [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lazy-loading 'pci_requests' on Instance uuid 27b14750-e349-47af-affe-02d10050aad1 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1171.611659] env[62569]: DEBUG nova.objects.base [None req-63d1e0c6-b93e-4aa1-9912-8b422733ffe4 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Object Instance<4689e0e5-1566-4ec7-8060-ed89c5db6919> lazy-loaded attributes: flavor,info_cache {{(pid=62569) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1171.826019] env[62569]: DEBUG nova.objects.instance [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lazy-loading 'numa_topology' on Instance uuid 27b14750-e349-47af-affe-02d10050aad1 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1172.317403] env[62569]: DEBUG nova.network.neutron [None req-63d1e0c6-b93e-4aa1-9912-8b422733ffe4 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Updating instance_info_cache with network_info: [{"id": "44c168a4-d010-4c11-946f-b85089f76985", "address": "fa:16:3e:d7:2e:b2", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44c168a4-d0", "ovs_interfaceid": "44c168a4-d010-4c11-946f-b85089f76985", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1172.328731] env[62569]: INFO nova.compute.claims [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1172.820804] env[62569]: DEBUG oslo_concurrency.lockutils [None req-63d1e0c6-b93e-4aa1-9912-8b422733ffe4 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "refresh_cache-4689e0e5-1566-4ec7-8060-ed89c5db6919" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1173.372030] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c9325ae-994f-4b1c-87ed-33b0faf8ea55 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.379459] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f62a3e6-8c35-443c-9cbd-7df54994fd20 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.408698] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84e0c91c-9651-4968-a582-1b504704d023 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.415378] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-022df7ba-508a-49da-86b4-6d30faa84cf6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.427970] env[62569]: DEBUG nova.compute.provider_tree [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1173.826650] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-63d1e0c6-b93e-4aa1-9912-8b422733ffe4 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1173.826983] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e1988cec-7c38-4cd8-8284-9b30809326c6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.833902] env[62569]: DEBUG oslo_vmware.api [None req-63d1e0c6-b93e-4aa1-9912-8b422733ffe4 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1173.833902] env[62569]: value = "task-1250763" [ 1173.833902] env[62569]: _type = "Task" [ 1173.833902] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.840914] env[62569]: DEBUG oslo_vmware.api [None req-63d1e0c6-b93e-4aa1-9912-8b422733ffe4 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250763, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.931811] env[62569]: DEBUG nova.scheduler.client.report [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1174.344203] env[62569]: DEBUG oslo_vmware.api [None req-63d1e0c6-b93e-4aa1-9912-8b422733ffe4 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250763, 'name': PowerOnVM_Task, 'duration_secs': 0.380965} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.344485] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-63d1e0c6-b93e-4aa1-9912-8b422733ffe4 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1174.344704] env[62569]: DEBUG nova.compute.manager [None req-63d1e0c6-b93e-4aa1-9912-8b422733ffe4 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1174.345520] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a165ca8-2d92-4460-9009-be879c4cc09b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.436520] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.115s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1174.464696] env[62569]: INFO nova.network.neutron [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Updating port 26f272f3-4864-483c-a76f-3bcf9efa4ab8 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1175.703929] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa4f4d9f-90b3-4cf5-93a9-eb10a32e6411 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.710424] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e46c1218-7bd2-4e69-bb2c-da38b8ded12c tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Suspending the VM {{(pid=62569) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1175.710662] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-4c8b55ea-6187-42f9-b2c0-3e2acd9a8867 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.717456] env[62569]: DEBUG oslo_vmware.api [None req-e46c1218-7bd2-4e69-bb2c-da38b8ded12c tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1175.717456] env[62569]: value = "task-1250764" [ 1175.717456] env[62569]: _type = "Task" [ 1175.717456] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.725051] env[62569]: DEBUG oslo_vmware.api [None req-e46c1218-7bd2-4e69-bb2c-da38b8ded12c tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250764, 'name': SuspendVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.077531] env[62569]: DEBUG nova.compute.manager [req-d76e81ee-3ec1-45ea-b086-0df1e51b7123 req-58c2600c-c108-41b4-848f-666c5e7d8b18 service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Received event network-vif-plugged-26f272f3-4864-483c-a76f-3bcf9efa4ab8 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1176.077772] env[62569]: DEBUG oslo_concurrency.lockutils [req-d76e81ee-3ec1-45ea-b086-0df1e51b7123 req-58c2600c-c108-41b4-848f-666c5e7d8b18 service nova] Acquiring lock "27b14750-e349-47af-affe-02d10050aad1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1176.077946] env[62569]: DEBUG oslo_concurrency.lockutils [req-d76e81ee-3ec1-45ea-b086-0df1e51b7123 req-58c2600c-c108-41b4-848f-666c5e7d8b18 service nova] Lock "27b14750-e349-47af-affe-02d10050aad1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1176.078164] env[62569]: DEBUG oslo_concurrency.lockutils [req-d76e81ee-3ec1-45ea-b086-0df1e51b7123 req-58c2600c-c108-41b4-848f-666c5e7d8b18 service nova] Lock "27b14750-e349-47af-affe-02d10050aad1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1176.078306] env[62569]: DEBUG nova.compute.manager [req-d76e81ee-3ec1-45ea-b086-0df1e51b7123 req-58c2600c-c108-41b4-848f-666c5e7d8b18 service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] No waiting events found dispatching network-vif-plugged-26f272f3-4864-483c-a76f-3bcf9efa4ab8 {{(pid=62569) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1176.078522] env[62569]: WARNING nova.compute.manager [req-d76e81ee-3ec1-45ea-b086-0df1e51b7123 req-58c2600c-c108-41b4-848f-666c5e7d8b18 service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Received unexpected event network-vif-plugged-26f272f3-4864-483c-a76f-3bcf9efa4ab8 for instance with vm_state shelved_offloaded and task_state spawning. [ 1176.165271] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1176.168107] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1176.168301] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.168481] env[62569]: DEBUG nova.network.neutron [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1176.227235] env[62569]: DEBUG oslo_vmware.api [None req-e46c1218-7bd2-4e69-bb2c-da38b8ded12c tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250764, 'name': SuspendVM_Task} progress is 70%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.655903] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1176.727567] env[62569]: DEBUG oslo_vmware.api [None req-e46c1218-7bd2-4e69-bb2c-da38b8ded12c tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250764, 'name': SuspendVM_Task, 'duration_secs': 0.615336} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.727938] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-e46c1218-7bd2-4e69-bb2c-da38b8ded12c tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Suspended the VM {{(pid=62569) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1176.727938] env[62569]: DEBUG nova.compute.manager [None req-e46c1218-7bd2-4e69-bb2c-da38b8ded12c tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1176.729839] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2cc7919-7d61-47a1-b884-b2ad8edecca1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.877523] env[62569]: DEBUG nova.network.neutron [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Updating instance_info_cache with network_info: [{"id": "26f272f3-4864-483c-a76f-3bcf9efa4ab8", "address": "fa:16:3e:c8:1f:5b", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26f272f3-48", "ovs_interfaceid": "26f272f3-4864-483c-a76f-3bcf9efa4ab8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1177.380389] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Releasing lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1177.409040] env[62569]: DEBUG nova.virt.hardware [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='16ed3d1c397932394cfa24de01c3d211',container_format='bare',created_at=2024-10-10T15:42:36Z,direct_url=,disk_format='vmdk',id=7d09e0bb-738b-4c70-8d49-3e9198c8c1f2,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-85831029-shelved',owner='00bdb4c3f88f4d61834f7e961629ef3a',properties=ImageMetaProps,protected=,size=31670272,status='active',tags=,updated_at=2024-10-10T15:42:50Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1177.409316] env[62569]: DEBUG nova.virt.hardware [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1177.409480] env[62569]: DEBUG nova.virt.hardware [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1177.409668] env[62569]: DEBUG nova.virt.hardware [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1177.409822] env[62569]: DEBUG nova.virt.hardware [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1177.409976] env[62569]: DEBUG nova.virt.hardware [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1177.410207] env[62569]: DEBUG nova.virt.hardware [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1177.410377] env[62569]: DEBUG nova.virt.hardware [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1177.410549] env[62569]: DEBUG nova.virt.hardware [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1177.410717] env[62569]: DEBUG nova.virt.hardware [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1177.410894] env[62569]: DEBUG nova.virt.hardware [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1177.411776] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2cecd1d-cc6e-4ffb-92fb-69b4341eb8fc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.419603] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6c48a46-5264-47f0-9e6d-9153e04c2063 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.434833] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c8:1f:5b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b356db78-99c7-4464-822c-fc7e193f7878', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '26f272f3-4864-483c-a76f-3bcf9efa4ab8', 'vif_model': 'vmxnet3'}] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1177.442218] env[62569]: DEBUG oslo.service.loopingcall [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1177.442476] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27b14750-e349-47af-affe-02d10050aad1] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1177.442846] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-085e71e2-5f86-41d2-9150-0c0b02e53f07 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.460822] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1177.460822] env[62569]: value = "task-1250765" [ 1177.460822] env[62569]: _type = "Task" [ 1177.460822] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.468160] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250765, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.970642] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250765, 'name': CreateVM_Task} progress is 99%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.050593] env[62569]: INFO nova.compute.manager [None req-cd5641dc-60b9-4d5f-83c6-018777f3893f tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Resuming [ 1178.051222] env[62569]: DEBUG nova.objects.instance [None req-cd5641dc-60b9-4d5f-83c6-018777f3893f tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lazy-loading 'flavor' on Instance uuid 4689e0e5-1566-4ec7-8060-ed89c5db6919 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1178.104379] env[62569]: DEBUG nova.compute.manager [req-925ada43-ef0d-45d0-b5cf-065f879d8f3f req-8e4937a0-7e68-4a95-9703-4709252c9f43 service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Received event network-changed-26f272f3-4864-483c-a76f-3bcf9efa4ab8 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1178.104579] env[62569]: DEBUG nova.compute.manager [req-925ada43-ef0d-45d0-b5cf-065f879d8f3f req-8e4937a0-7e68-4a95-9703-4709252c9f43 service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Refreshing instance network info cache due to event network-changed-26f272f3-4864-483c-a76f-3bcf9efa4ab8. {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11516}} [ 1178.104794] env[62569]: DEBUG oslo_concurrency.lockutils [req-925ada43-ef0d-45d0-b5cf-065f879d8f3f req-8e4937a0-7e68-4a95-9703-4709252c9f43 service nova] Acquiring lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1178.104940] env[62569]: DEBUG oslo_concurrency.lockutils [req-925ada43-ef0d-45d0-b5cf-065f879d8f3f req-8e4937a0-7e68-4a95-9703-4709252c9f43 service nova] Acquired lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.105154] env[62569]: DEBUG nova.network.neutron [req-925ada43-ef0d-45d0-b5cf-065f879d8f3f req-8e4937a0-7e68-4a95-9703-4709252c9f43 service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Refreshing network info cache for port 26f272f3-4864-483c-a76f-3bcf9efa4ab8 {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1178.471087] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250765, 'name': CreateVM_Task, 'duration_secs': 0.544161} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.471280] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27b14750-e349-47af-affe-02d10050aad1] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1178.471920] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7d09e0bb-738b-4c70-8d49-3e9198c8c1f2" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1178.472108] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7d09e0bb-738b-4c70-8d49-3e9198c8c1f2" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.472502] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7d09e0bb-738b-4c70-8d49-3e9198c8c1f2" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1178.472762] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acb352b5-d469-4c1e-a9bd-37e916272e58 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.476810] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1178.476810] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529646fe-27c1-de10-89d1-1745542df032" [ 1178.476810] env[62569]: _type = "Task" [ 1178.476810] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.483791] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]529646fe-27c1-de10-89d1-1745542df032, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.864576] env[62569]: DEBUG nova.network.neutron [req-925ada43-ef0d-45d0-b5cf-065f879d8f3f req-8e4937a0-7e68-4a95-9703-4709252c9f43 service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Updated VIF entry in instance network info cache for port 26f272f3-4864-483c-a76f-3bcf9efa4ab8. {{(pid=62569) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1178.865071] env[62569]: DEBUG nova.network.neutron [req-925ada43-ef0d-45d0-b5cf-065f879d8f3f req-8e4937a0-7e68-4a95-9703-4709252c9f43 service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Updating instance_info_cache with network_info: [{"id": "26f272f3-4864-483c-a76f-3bcf9efa4ab8", "address": "fa:16:3e:c8:1f:5b", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26f272f3-48", "ovs_interfaceid": "26f272f3-4864-483c-a76f-3bcf9efa4ab8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1178.986477] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7d09e0bb-738b-4c70-8d49-3e9198c8c1f2" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1178.986887] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Processing image 7d09e0bb-738b-4c70-8d49-3e9198c8c1f2 {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1178.986952] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7d09e0bb-738b-4c70-8d49-3e9198c8c1f2/7d09e0bb-738b-4c70-8d49-3e9198c8c1f2.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1178.987123] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7d09e0bb-738b-4c70-8d49-3e9198c8c1f2/7d09e0bb-738b-4c70-8d49-3e9198c8c1f2.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.987309] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1178.987562] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f32349f6-1a20-4413-8d78-f8652fbffc7a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.995258] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1178.995431] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1178.996102] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-722200d3-347d-427c-a7aa-9b45a5499a44 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.000683] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1179.000683] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52dbc0bd-e581-3751-2682-78db4c262ff1" [ 1179.000683] env[62569]: _type = "Task" [ 1179.000683] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.007560] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52dbc0bd-e581-3751-2682-78db4c262ff1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.367616] env[62569]: DEBUG oslo_concurrency.lockutils [req-925ada43-ef0d-45d0-b5cf-065f879d8f3f req-8e4937a0-7e68-4a95-9703-4709252c9f43 service nova] Releasing lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1179.512644] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Preparing fetch location {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1179.512948] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Fetch image to [datastore2] OSTACK_IMG_a80d1e9a-efef-4e1e-8d83-9b19ff37964b/OSTACK_IMG_a80d1e9a-efef-4e1e-8d83-9b19ff37964b.vmdk {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1179.513161] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Downloading stream optimized image 7d09e0bb-738b-4c70-8d49-3e9198c8c1f2 to [datastore2] OSTACK_IMG_a80d1e9a-efef-4e1e-8d83-9b19ff37964b/OSTACK_IMG_a80d1e9a-efef-4e1e-8d83-9b19ff37964b.vmdk on the data store datastore2 as vApp {{(pid=62569) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1179.513341] env[62569]: DEBUG nova.virt.vmwareapi.images [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Downloading image file data 7d09e0bb-738b-4c70-8d49-3e9198c8c1f2 to the ESX as VM named 'OSTACK_IMG_a80d1e9a-efef-4e1e-8d83-9b19ff37964b' {{(pid=62569) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1179.560482] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cd5641dc-60b9-4d5f-83c6-018777f3893f tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "refresh_cache-4689e0e5-1566-4ec7-8060-ed89c5db6919" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1179.560670] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cd5641dc-60b9-4d5f-83c6-018777f3893f tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquired lock "refresh_cache-4689e0e5-1566-4ec7-8060-ed89c5db6919" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1179.560845] env[62569]: DEBUG nova.network.neutron [None req-cd5641dc-60b9-4d5f-83c6-018777f3893f tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1179.580813] env[62569]: DEBUG oslo_vmware.rw_handles [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1179.580813] env[62569]: value = "resgroup-9" [ 1179.580813] env[62569]: _type = "ResourcePool" [ 1179.580813] env[62569]: }. {{(pid=62569) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1179.581098] env[62569]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-95a9202a-7286-4f1a-a02e-e746ecbd1a77 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.601197] env[62569]: DEBUG oslo_vmware.rw_handles [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lease: (returnval){ [ 1179.601197] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5220c8e3-7c1f-4e88-44fd-b00294e0e1f4" [ 1179.601197] env[62569]: _type = "HttpNfcLease" [ 1179.601197] env[62569]: } obtained for vApp import into resource pool (val){ [ 1179.601197] env[62569]: value = "resgroup-9" [ 1179.601197] env[62569]: _type = "ResourcePool" [ 1179.601197] env[62569]: }. {{(pid=62569) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1179.601465] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the lease: (returnval){ [ 1179.601465] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5220c8e3-7c1f-4e88-44fd-b00294e0e1f4" [ 1179.601465] env[62569]: _type = "HttpNfcLease" [ 1179.601465] env[62569]: } to be ready. {{(pid=62569) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1179.607259] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1179.607259] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5220c8e3-7c1f-4e88-44fd-b00294e0e1f4" [ 1179.607259] env[62569]: _type = "HttpNfcLease" [ 1179.607259] env[62569]: } is initializing. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1179.656560] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1180.108677] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1180.108677] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5220c8e3-7c1f-4e88-44fd-b00294e0e1f4" [ 1180.108677] env[62569]: _type = "HttpNfcLease" [ 1180.108677] env[62569]: } is initializing. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1180.258884] env[62569]: DEBUG nova.network.neutron [None req-cd5641dc-60b9-4d5f-83c6-018777f3893f tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Updating instance_info_cache with network_info: [{"id": "44c168a4-d010-4c11-946f-b85089f76985", "address": "fa:16:3e:d7:2e:b2", "network": {"id": "373cdd7d-542a-4782-8b14-41caa5690188", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1089740252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f988acdb9cd4ca28ca8916c2ede2db5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73915082-a1b0-460b-b24d-97588fc9cb29", "external-id": "nsx-vlan-transportzone-744", "segmentation_id": 744, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44c168a4-d0", "ovs_interfaceid": "44c168a4-d010-4c11-946f-b85089f76985", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1180.609827] env[62569]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1180.609827] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5220c8e3-7c1f-4e88-44fd-b00294e0e1f4" [ 1180.609827] env[62569]: _type = "HttpNfcLease" [ 1180.609827] env[62569]: } is ready. {{(pid=62569) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1180.610152] env[62569]: DEBUG oslo_vmware.rw_handles [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1180.610152] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5220c8e3-7c1f-4e88-44fd-b00294e0e1f4" [ 1180.610152] env[62569]: _type = "HttpNfcLease" [ 1180.610152] env[62569]: }. {{(pid=62569) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1180.610879] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-672f10e7-fed7-457c-af22-ef6eca91a069 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.617892] env[62569]: DEBUG oslo_vmware.rw_handles [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52334f69-fdd3-f270-85ab-8f431b2fa234/disk-0.vmdk from lease info. {{(pid=62569) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1180.618087] env[62569]: DEBUG oslo_vmware.rw_handles [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Creating HTTP connection to write to file with size = 31670272 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52334f69-fdd3-f270-85ab-8f431b2fa234/disk-0.vmdk. {{(pid=62569) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1180.673584] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1180.673772] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1180.673916] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62569) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}} [ 1180.681716] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-aefb6478-caf3-4498-a64c-00c6044af259 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.761859] env[62569]: DEBUG oslo_concurrency.lockutils [None req-cd5641dc-60b9-4d5f-83c6-018777f3893f tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Releasing lock "refresh_cache-4689e0e5-1566-4ec7-8060-ed89c5db6919" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1180.762951] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-489d2297-46bb-4672-b3d9-a51fdc3d2cca {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.769889] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cd5641dc-60b9-4d5f-83c6-018777f3893f tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Resuming the VM {{(pid=62569) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1180.770157] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-034db64d-c72f-40f0-8425-0fdd130296c2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.775574] env[62569]: DEBUG oslo_vmware.api [None req-cd5641dc-60b9-4d5f-83c6-018777f3893f tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1180.775574] env[62569]: value = "task-1250767" [ 1180.775574] env[62569]: _type = "Task" [ 1180.775574] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.783527] env[62569]: DEBUG oslo_vmware.api [None req-cd5641dc-60b9-4d5f-83c6-018777f3893f tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250767, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.287049] env[62569]: DEBUG oslo_vmware.api [None req-cd5641dc-60b9-4d5f-83c6-018777f3893f tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250767, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.787783] env[62569]: DEBUG oslo_vmware.api [None req-cd5641dc-60b9-4d5f-83c6-018777f3893f tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250767, 'name': PowerOnVM_Task, 'duration_secs': 0.545932} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.789355] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-cd5641dc-60b9-4d5f-83c6-018777f3893f tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Resumed the VM {{(pid=62569) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1181.789560] env[62569]: DEBUG nova.compute.manager [None req-cd5641dc-60b9-4d5f-83c6-018777f3893f tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1181.790447] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf0c8002-74f1-44f0-967b-582795782193 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.048235] env[62569]: DEBUG oslo_vmware.rw_handles [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Completed reading data from the image iterator. {{(pid=62569) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1182.048574] env[62569]: DEBUG oslo_vmware.rw_handles [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52334f69-fdd3-f270-85ab-8f431b2fa234/disk-0.vmdk. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1182.049686] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea02333f-ca0c-4bea-a171-44e131937765 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.057521] env[62569]: DEBUG oslo_vmware.rw_handles [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52334f69-fdd3-f270-85ab-8f431b2fa234/disk-0.vmdk is in state: ready. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1182.057750] env[62569]: DEBUG oslo_vmware.rw_handles [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52334f69-fdd3-f270-85ab-8f431b2fa234/disk-0.vmdk. {{(pid=62569) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1182.057991] env[62569]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-1d17d121-cfa6-485f-9945-67459964b6cd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.656590] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1182.656945] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Starting heal instance info cache {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10308}} [ 1182.656945] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Rebuilding the list of instances to heal {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10312}} [ 1182.817870] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "4689e0e5-1566-4ec7-8060-ed89c5db6919" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1182.818142] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "4689e0e5-1566-4ec7-8060-ed89c5db6919" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1182.818366] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "4689e0e5-1566-4ec7-8060-ed89c5db6919-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1182.818553] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "4689e0e5-1566-4ec7-8060-ed89c5db6919-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1182.818725] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "4689e0e5-1566-4ec7-8060-ed89c5db6919-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1182.820751] env[62569]: INFO nova.compute.manager [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Terminating instance [ 1182.904734] env[62569]: DEBUG oslo_vmware.rw_handles [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52334f69-fdd3-f270-85ab-8f431b2fa234/disk-0.vmdk. {{(pid=62569) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1182.904970] env[62569]: INFO nova.virt.vmwareapi.images [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Downloaded image file data 7d09e0bb-738b-4c70-8d49-3e9198c8c1f2 [ 1182.905862] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb1bee81-645d-4279-adcf-7cd4128c1367 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.921138] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7be2aa0e-9688-44e5-a2bf-da8a5cbf770d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.973373] env[62569]: INFO nova.virt.vmwareapi.images [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] The imported VM was unregistered [ 1182.975699] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Caching image {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1182.975934] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Creating directory with path [datastore2] devstack-image-cache_base/7d09e0bb-738b-4c70-8d49-3e9198c8c1f2 {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1182.976214] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-81f6abd5-2883-4e2a-99ba-ea1fae0bf1bc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.996540] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Created directory with path [datastore2] devstack-image-cache_base/7d09e0bb-738b-4c70-8d49-3e9198c8c1f2 {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1182.996729] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_a80d1e9a-efef-4e1e-8d83-9b19ff37964b/OSTACK_IMG_a80d1e9a-efef-4e1e-8d83-9b19ff37964b.vmdk to [datastore2] devstack-image-cache_base/7d09e0bb-738b-4c70-8d49-3e9198c8c1f2/7d09e0bb-738b-4c70-8d49-3e9198c8c1f2.vmdk. {{(pid=62569) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1182.996971] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-ff50dd21-bfed-45b9-990e-70c20000d77c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.003164] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1183.003164] env[62569]: value = "task-1250769" [ 1183.003164] env[62569]: _type = "Task" [ 1183.003164] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.010294] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250769, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.162384] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Skipping network cache update for instance because it is being deleted. {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10325}} [ 1183.162631] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1183.162755] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquired lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.162896] env[62569]: DEBUG nova.network.neutron [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 27b14750-e349-47af-affe-02d10050aad1] Forcefully refreshing network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1183.163075] env[62569]: DEBUG nova.objects.instance [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lazy-loading 'info_cache' on Instance uuid 27b14750-e349-47af-affe-02d10050aad1 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1183.324417] env[62569]: DEBUG nova.compute.manager [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1183.324703] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1183.325690] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-082bfb6a-5783-4a47-a2ef-516048d56a30 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.336829] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1183.337207] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-99ceca8c-7d8c-4208-a698-355364c1f184 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.348217] env[62569]: DEBUG oslo_vmware.api [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1183.348217] env[62569]: value = "task-1250770" [ 1183.348217] env[62569]: _type = "Task" [ 1183.348217] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.363155] env[62569]: DEBUG oslo_vmware.api [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250770, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.518033] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250769, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.860181] env[62569]: DEBUG oslo_vmware.api [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250770, 'name': PowerOffVM_Task, 'duration_secs': 0.213682} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.860540] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1183.860662] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1183.860895] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eb5fd081-7a1f-40de-a414-2d9d2b614fa0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.928799] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1183.929046] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1183.929242] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Deleting the datastore file [datastore1] 4689e0e5-1566-4ec7-8060-ed89c5db6919 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1183.929520] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-846c15b8-febf-4263-84a3-fc8714ffdf50 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.939352] env[62569]: DEBUG oslo_vmware.api [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for the task: (returnval){ [ 1183.939352] env[62569]: value = "task-1250772" [ 1183.939352] env[62569]: _type = "Task" [ 1183.939352] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.952087] env[62569]: DEBUG oslo_vmware.api [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250772, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.017377] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250769, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.453426] env[62569]: DEBUG oslo_vmware.api [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Task: {'id': task-1250772, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.233863} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.453768] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1184.454012] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1184.454215] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1184.454399] env[62569]: INFO nova.compute.manager [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1184.454651] env[62569]: DEBUG oslo.service.loopingcall [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1184.454892] env[62569]: DEBUG nova.compute.manager [-] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1184.455036] env[62569]: DEBUG nova.network.neutron [-] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1184.518518] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250769, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.893883] env[62569]: DEBUG nova.network.neutron [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 27b14750-e349-47af-affe-02d10050aad1] Updating instance_info_cache with network_info: [{"id": "26f272f3-4864-483c-a76f-3bcf9efa4ab8", "address": "fa:16:3e:c8:1f:5b", "network": {"id": "b1b5a18f-bf4f-4410-af6b-bfa98b77bb40", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-2030081341-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "00bdb4c3f88f4d61834f7e961629ef3a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26f272f3-48", "ovs_interfaceid": "26f272f3-4864-483c-a76f-3bcf9efa4ab8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1185.019502] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250769, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.067391] env[62569]: DEBUG nova.compute.manager [req-f0cd84c9-b84e-4bdd-b6d4-f03d7b38e71c req-f8138ed6-18cd-4401-b0d4-a0a26a821d88 service nova] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Received event network-vif-deleted-44c168a4-d010-4c11-946f-b85089f76985 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1185.067668] env[62569]: INFO nova.compute.manager [req-f0cd84c9-b84e-4bdd-b6d4-f03d7b38e71c req-f8138ed6-18cd-4401-b0d4-a0a26a821d88 service nova] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Neutron deleted interface 44c168a4-d010-4c11-946f-b85089f76985; detaching it from the instance and deleting it from the info cache [ 1185.067895] env[62569]: DEBUG nova.network.neutron [req-f0cd84c9-b84e-4bdd-b6d4-f03d7b38e71c req-f8138ed6-18cd-4401-b0d4-a0a26a821d88 service nova] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1185.398058] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Releasing lock "refresh_cache-27b14750-e349-47af-affe-02d10050aad1" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1185.398058] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] [instance: 27b14750-e349-47af-affe-02d10050aad1] Updated the network info_cache for instance {{(pid=62569) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10379}} [ 1185.398058] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1185.398058] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1185.398058] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1185.515757] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250769, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.2343} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.516037] env[62569]: INFO nova.virt.vmwareapi.ds_util [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_a80d1e9a-efef-4e1e-8d83-9b19ff37964b/OSTACK_IMG_a80d1e9a-efef-4e1e-8d83-9b19ff37964b.vmdk to [datastore2] devstack-image-cache_base/7d09e0bb-738b-4c70-8d49-3e9198c8c1f2/7d09e0bb-738b-4c70-8d49-3e9198c8c1f2.vmdk. [ 1185.516274] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Cleaning up location [datastore2] OSTACK_IMG_a80d1e9a-efef-4e1e-8d83-9b19ff37964b {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1185.516464] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_a80d1e9a-efef-4e1e-8d83-9b19ff37964b {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1185.516714] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f5992e9a-c12d-4f41-b4bf-17c99a63ee8c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.523206] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1185.523206] env[62569]: value = "task-1250773" [ 1185.523206] env[62569]: _type = "Task" [ 1185.523206] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.530535] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250773, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.538182] env[62569]: DEBUG nova.network.neutron [-] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1185.570486] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-87e4a998-8322-4dc8-a663-5f5f5b702241 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.579649] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bed8fbbd-6da8-4eba-ac1f-73d5aea0b0b1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.602050] env[62569]: DEBUG nova.compute.manager [req-f0cd84c9-b84e-4bdd-b6d4-f03d7b38e71c req-f8138ed6-18cd-4401-b0d4-a0a26a821d88 service nova] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Detach interface failed, port_id=44c168a4-d010-4c11-946f-b85089f76985, reason: Instance 4689e0e5-1566-4ec7-8060-ed89c5db6919 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1185.901021] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.901338] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.901419] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.901576] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62569) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1185.902508] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdcf0bf5-072c-4abd-b3a1-ca735bcabe6b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.910706] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4b7d353-8912-4593-90d4-c04afc9d5161 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.925055] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf8b631f-47e4-47f4-8a13-b3d58e15df80 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.930889] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9258809f-6b23-4da2-bcfe-bfef872169db {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.958412] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181174MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=62569) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1185.958544] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.958728] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1186.032816] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250773, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.063307} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.034022] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1186.034022] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7d09e0bb-738b-4c70-8d49-3e9198c8c1f2/7d09e0bb-738b-4c70-8d49-3e9198c8c1f2.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1186.034022] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7d09e0bb-738b-4c70-8d49-3e9198c8c1f2/7d09e0bb-738b-4c70-8d49-3e9198c8c1f2.vmdk to [datastore2] 27b14750-e349-47af-affe-02d10050aad1/27b14750-e349-47af-affe-02d10050aad1.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1186.034022] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7d4ee058-d8c7-4d83-960c-1e5599c67ee2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.040555] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1186.040555] env[62569]: value = "task-1250774" [ 1186.040555] env[62569]: _type = "Task" [ 1186.040555] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.040908] env[62569]: INFO nova.compute.manager [-] [instance: 4689e0e5-1566-4ec7-8060-ed89c5db6919] Took 1.59 seconds to deallocate network for instance. [ 1186.052914] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250774, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.553499] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1186.553810] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250774, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.987232] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 4689e0e5-1566-4ec7-8060-ed89c5db6919 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1186.987507] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Instance 27b14750-e349-47af-affe-02d10050aad1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62569) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1186.987685] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62569) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1186.987899] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62569) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1187.030129] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1802ecc2-074d-47f7-93a0-22b37afa8a96 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.042376] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4b1c90b-e144-4818-a141-281b506d53e1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.054946] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250774, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.082834] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b42a96d3-3713-4164-a8fd-021ddceeec5e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.090634] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a450542c-f645-4265-abd9-e0d1f394d16e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.109414] env[62569]: DEBUG nova.compute.provider_tree [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1187.557347] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250774, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.612604] env[62569]: DEBUG nova.scheduler.client.report [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1188.059707] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250774, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.117868] env[62569]: DEBUG nova.compute.resource_tracker [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62569) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1188.118090] env[62569]: DEBUG oslo_concurrency.lockutils [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.159s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1188.118435] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.565s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.118681] env[62569]: DEBUG nova.objects.instance [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lazy-loading 'resources' on Instance uuid 4689e0e5-1566-4ec7-8060-ed89c5db6919 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1188.557295] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250774, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.209572} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.557569] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7d09e0bb-738b-4c70-8d49-3e9198c8c1f2/7d09e0bb-738b-4c70-8d49-3e9198c8c1f2.vmdk to [datastore2] 27b14750-e349-47af-affe-02d10050aad1/27b14750-e349-47af-affe-02d10050aad1.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1188.558355] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2b32528-7c08-44b9-a573-25098e5518e2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.579045] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 27b14750-e349-47af-affe-02d10050aad1/27b14750-e349-47af-affe-02d10050aad1.vmdk or device None with type streamOptimized {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1188.579273] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2454eca3-51a8-4bcc-87f5-58d92e054c8a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.598745] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1188.598745] env[62569]: value = "task-1250775" [ 1188.598745] env[62569]: _type = "Task" [ 1188.598745] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.605608] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250775, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.656284] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51652138-9419-4f18-b5fc-44a9232e4bac {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.663158] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35a360d6-a373-45f7-a708-ec7a2f7d5399 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.692240] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e20ca520-9eda-4f72-8ff4-86d0382af1ce {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.698899] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-606b707e-cb5c-4fa1-9b54-5a875056dddc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.711677] env[62569]: DEBUG nova.compute.provider_tree [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1189.108611] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250775, 'name': ReconfigVM_Task, 'duration_secs': 0.331588} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.108953] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 27b14750-e349-47af-affe-02d10050aad1/27b14750-e349-47af-affe-02d10050aad1.vmdk or device None with type streamOptimized {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1189.109610] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-38e6888f-e1e1-4e57-a7c4-daa9c4181ec8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.115086] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1189.115086] env[62569]: value = "task-1250776" [ 1189.115086] env[62569]: _type = "Task" [ 1189.115086] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.122225] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250776, 'name': Rename_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.214987] env[62569]: DEBUG nova.scheduler.client.report [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1189.625454] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250776, 'name': Rename_Task, 'duration_secs': 0.132778} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.625746] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1189.625988] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-93043631-2c11-487e-a2fc-ad4f6c6a3ce6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.632327] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1189.632327] env[62569]: value = "task-1250777" [ 1189.632327] env[62569]: _type = "Task" [ 1189.632327] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.640282] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250777, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.720448] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.602s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1189.741491] env[62569]: INFO nova.scheduler.client.report [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Deleted allocations for instance 4689e0e5-1566-4ec7-8060-ed89c5db6919 [ 1190.142627] env[62569]: DEBUG oslo_vmware.api [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250777, 'name': PowerOnVM_Task, 'duration_secs': 0.47469} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.142627] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1190.244265] env[62569]: DEBUG nova.compute.manager [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1190.245103] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db5852a0-79fe-410e-a737-b7cd120f2126 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.250655] env[62569]: DEBUG oslo_concurrency.lockutils [None req-7a16f328-458c-4007-b6d8-02daac61a2f5 tempest-ServerActionsTestJSON-1480668520 tempest-ServerActionsTestJSON-1480668520-project-member] Lock "4689e0e5-1566-4ec7-8060-ed89c5db6919" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.432s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.762049] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f53fad9b-32be-4a62-95a4-02adaabdac0d tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "27b14750-e349-47af-affe-02d10050aad1" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 20.468s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.277113] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "27b14750-e349-47af-affe-02d10050aad1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.277113] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "27b14750-e349-47af-affe-02d10050aad1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.277113] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "27b14750-e349-47af-affe-02d10050aad1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.277457] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "27b14750-e349-47af-affe-02d10050aad1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.277457] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "27b14750-e349-47af-affe-02d10050aad1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.281922] env[62569]: INFO nova.compute.manager [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Terminating instance [ 1191.788321] env[62569]: DEBUG nova.compute.manager [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1191.788565] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1191.789504] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5516cab-1f8d-4737-a940-37c6aa15d04a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.797877] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1191.797877] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f66f31a-81ee-41a7-8980-2a7ed71fac0e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.803877] env[62569]: DEBUG oslo_vmware.api [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1191.803877] env[62569]: value = "task-1250778" [ 1191.803877] env[62569]: _type = "Task" [ 1191.803877] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.812546] env[62569]: DEBUG oslo_vmware.api [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250778, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.314014] env[62569]: DEBUG oslo_vmware.api [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250778, 'name': PowerOffVM_Task, 'duration_secs': 0.199742} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.314339] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1192.314506] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1192.314754] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0f513193-06e8-4dd4-9e78-e66fb9285319 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.374924] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1192.375188] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Deleting contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1192.375450] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Deleting the datastore file [datastore2] 27b14750-e349-47af-affe-02d10050aad1 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1192.375730] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f8ce539a-6496-4316-bdd6-61c8bead95e8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.382608] env[62569]: DEBUG oslo_vmware.api [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for the task: (returnval){ [ 1192.382608] env[62569]: value = "task-1250780" [ 1192.382608] env[62569]: _type = "Task" [ 1192.382608] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.390198] env[62569]: DEBUG oslo_vmware.api [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250780, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.892971] env[62569]: DEBUG oslo_vmware.api [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Task: {'id': task-1250780, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.122501} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.893187] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1192.893417] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Deleted contents of the VM from datastore datastore2 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1192.893602] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1192.893785] env[62569]: INFO nova.compute.manager [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] [instance: 27b14750-e349-47af-affe-02d10050aad1] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1192.894037] env[62569]: DEBUG oslo.service.loopingcall [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1192.894250] env[62569]: DEBUG nova.compute.manager [-] [instance: 27b14750-e349-47af-affe-02d10050aad1] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1192.894341] env[62569]: DEBUG nova.network.neutron [-] [instance: 27b14750-e349-47af-affe-02d10050aad1] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1193.109927] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Acquiring lock "3e0b6e07-b175-403d-9e53-c52e34a1c507" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1193.110179] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Lock "3e0b6e07-b175-403d-9e53-c52e34a1c507" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1193.610241] env[62569]: DEBUG nova.compute.manager [req-c6286cf6-a76b-4d3c-aded-7487faaae36d req-7c233d2d-6525-4adb-ade9-b61424f18321 service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Received event network-vif-deleted-26f272f3-4864-483c-a76f-3bcf9efa4ab8 {{(pid=62569) external_instance_event /opt/stack/nova/nova/compute/manager.py:11511}} [ 1193.610528] env[62569]: INFO nova.compute.manager [req-c6286cf6-a76b-4d3c-aded-7487faaae36d req-7c233d2d-6525-4adb-ade9-b61424f18321 service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Neutron deleted interface 26f272f3-4864-483c-a76f-3bcf9efa4ab8; detaching it from the instance and deleting it from the info cache [ 1193.610615] env[62569]: DEBUG nova.network.neutron [req-c6286cf6-a76b-4d3c-aded-7487faaae36d req-7c233d2d-6525-4adb-ade9-b61424f18321 service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1193.616317] env[62569]: DEBUG nova.compute.manager [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Starting instance... {{(pid=62569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1194.081406] env[62569]: DEBUG nova.network.neutron [-] [instance: 27b14750-e349-47af-affe-02d10050aad1] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1194.114036] env[62569]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d09332e4-b024-4465-849a-67d7111881a2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.123388] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c62a98-d1a6-4894-aac7-3ad19502f423 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.149728] env[62569]: DEBUG nova.compute.manager [req-c6286cf6-a76b-4d3c-aded-7487faaae36d req-7c233d2d-6525-4adb-ade9-b61424f18321 service nova] [instance: 27b14750-e349-47af-affe-02d10050aad1] Detach interface failed, port_id=26f272f3-4864-483c-a76f-3bcf9efa4ab8, reason: Instance 27b14750-e349-47af-affe-02d10050aad1 could not be found. {{(pid=62569) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11345}} [ 1194.152249] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1194.152502] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1194.154029] env[62569]: INFO nova.compute.claims [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1194.582745] env[62569]: INFO nova.compute.manager [-] [instance: 27b14750-e349-47af-affe-02d10050aad1] Took 1.69 seconds to deallocate network for instance. [ 1195.090063] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1195.113716] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1195.199214] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-079f5d73-1304-41ae-aa72-7df3321db55b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.206696] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-256f0ddc-2e9d-44cb-ad3a-c8802f249b20 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.236510] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ec9d16-3d09-4846-89dd-be2cf1a1102d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.243327] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f7a8ff-dc89-48a4-b711-08c8d7f3f83f {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.255962] env[62569]: DEBUG nova.compute.provider_tree [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1195.759502] env[62569]: DEBUG nova.scheduler.client.report [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1196.264835] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.112s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1196.265413] env[62569]: DEBUG nova.compute.manager [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Start building networks asynchronously for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1196.268027] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.178s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.268615] env[62569]: DEBUG nova.objects.instance [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lazy-loading 'resources' on Instance uuid 27b14750-e349-47af-affe-02d10050aad1 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1196.773026] env[62569]: DEBUG nova.compute.utils [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Using /dev/sd instead of None {{(pid=62569) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1196.776283] env[62569]: DEBUG nova.compute.manager [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Not allocating networking since 'none' was specified. {{(pid=62569) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 1196.810516] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a41571-aa98-4f46-ab75-b63ca49ff3dc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.818171] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3fa82b4-b783-45e3-9b25-fc5800bc8040 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.848124] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70efeb6d-fcad-4daf-bec3-fec96c043ba0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.855455] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c08714-70a6-40de-8b30-6b58663c138c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.868479] env[62569]: DEBUG nova.compute.provider_tree [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1197.277633] env[62569]: DEBUG nova.compute.manager [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Start building block device mappings for instance. {{(pid=62569) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1197.371771] env[62569]: DEBUG nova.scheduler.client.report [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1197.876425] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.608s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.894378] env[62569]: INFO nova.scheduler.client.report [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Deleted allocations for instance 27b14750-e349-47af-affe-02d10050aad1 [ 1198.287304] env[62569]: DEBUG nova.compute.manager [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Start spawning the instance on the hypervisor. {{(pid=62569) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1198.312234] env[62569]: DEBUG nova.virt.hardware [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1198.312502] env[62569]: DEBUG nova.virt.hardware [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1198.312669] env[62569]: DEBUG nova.virt.hardware [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1198.312857] env[62569]: DEBUG nova.virt.hardware [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1198.313016] env[62569]: DEBUG nova.virt.hardware [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1198.313177] env[62569]: DEBUG nova.virt.hardware [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1198.313420] env[62569]: DEBUG nova.virt.hardware [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1198.313596] env[62569]: DEBUG nova.virt.hardware [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1198.313769] env[62569]: DEBUG nova.virt.hardware [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1198.313945] env[62569]: DEBUG nova.virt.hardware [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1198.314131] env[62569]: DEBUG nova.virt.hardware [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1198.314993] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a18fc322-2893-44d5-b1a5-ef9b6bbe2ea5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.322738] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa1d9ee2-e35a-4177-abc9-70df1a91481a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.335739] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Instance VIF info [] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1198.341144] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Creating folder: Project (628984db8a9441b5a0a8719f35b5c201). Parent ref: group-v269330. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1198.341402] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-01169062-f428-4705-9cff-63db8a1c24a8 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.350661] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Created folder: Project (628984db8a9441b5a0a8719f35b5c201) in parent group-v269330. [ 1198.350842] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Creating folder: Instances. Parent ref: group-v269517. {{(pid=62569) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1198.351049] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f80da508-5a6e-4cb2-afe1-b56cfd7804c5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.358542] env[62569]: INFO nova.virt.vmwareapi.vm_util [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Created folder: Instances in parent group-v269517. [ 1198.358763] env[62569]: DEBUG oslo.service.loopingcall [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1198.358944] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1198.359148] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5186875e-4d29-48c0-8f08-ef9542981a0c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.373979] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1198.373979] env[62569]: value = "task-1250783" [ 1198.373979] env[62569]: _type = "Task" [ 1198.373979] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.380487] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250783, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.401280] env[62569]: DEBUG oslo_concurrency.lockutils [None req-47644bcb-eef4-4314-95a1-bce8da4b2aa1 tempest-ServerActionsTestOtherB-1908205650 tempest-ServerActionsTestOtherB-1908205650-project-member] Lock "27b14750-e349-47af-affe-02d10050aad1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.124s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.885938] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250783, 'name': CreateVM_Task, 'duration_secs': 0.241432} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.886123] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1198.886554] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1198.886734] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1198.887067] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1198.887320] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac5c8656-60f5-4abd-8433-e60097e02983 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.891833] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Waiting for the task: (returnval){ [ 1198.891833] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52aff2d4-0073-8bc0-2d8b-8d4cce6c836a" [ 1198.891833] env[62569]: _type = "Task" [ 1198.891833] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.898693] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52aff2d4-0073-8bc0-2d8b-8d4cce6c836a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.402702] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52aff2d4-0073-8bc0-2d8b-8d4cce6c836a, 'name': SearchDatastore_Task, 'duration_secs': 0.009602} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.403196] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1199.403196] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1199.403491] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1199.403491] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.403630] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1199.403891] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-62a42f10-81a7-4f19-89a0-23ac5ca4b998 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.411439] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1199.411613] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1199.412291] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-def17ffe-9997-4316-b85c-1c75da3b20b0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.416955] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Waiting for the task: (returnval){ [ 1199.416955] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527113d3-7c01-8cde-f410-464760d471f2" [ 1199.416955] env[62569]: _type = "Task" [ 1199.416955] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.425637] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527113d3-7c01-8cde-f410-464760d471f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.927334] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527113d3-7c01-8cde-f410-464760d471f2, 'name': SearchDatastore_Task, 'duration_secs': 0.007511} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.928249] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6354939-eb17-477d-8fc6-885160602652 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.933468] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Waiting for the task: (returnval){ [ 1199.933468] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5227b536-52a8-a905-5bc2-cc703f8e6297" [ 1199.933468] env[62569]: _type = "Task" [ 1199.933468] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.940571] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5227b536-52a8-a905-5bc2-cc703f8e6297, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.443231] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]5227b536-52a8-a905-5bc2-cc703f8e6297, 'name': SearchDatastore_Task, 'duration_secs': 0.009328} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.443622] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1200.443875] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 3e0b6e07-b175-403d-9e53-c52e34a1c507/3e0b6e07-b175-403d-9e53-c52e34a1c507.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1200.444164] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6dd11e2a-537f-43a1-aaca-6f201f1137f6 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.451738] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Waiting for the task: (returnval){ [ 1200.451738] env[62569]: value = "task-1250785" [ 1200.451738] env[62569]: _type = "Task" [ 1200.451738] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.459353] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250785, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.961497] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250785, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.397627} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.961841] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 3e0b6e07-b175-403d-9e53-c52e34a1c507/3e0b6e07-b175-403d-9e53-c52e34a1c507.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1200.962111] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1200.962363] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-972b01c0-71e0-4f8c-a016-fab06b27165b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.968587] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Waiting for the task: (returnval){ [ 1200.968587] env[62569]: value = "task-1250786" [ 1200.968587] env[62569]: _type = "Task" [ 1200.968587] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.977519] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250786, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.478510] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250786, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066097} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.478879] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1201.479601] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b94d367-e19c-4139-a6c9-fa2956d5df58 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.498656] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Reconfiguring VM instance instance-00000071 to attach disk [datastore1] 3e0b6e07-b175-403d-9e53-c52e34a1c507/3e0b6e07-b175-403d-9e53-c52e34a1c507.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1201.498913] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2074b824-0cc6-4a32-ad2d-fc8dadc395a3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.518793] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Waiting for the task: (returnval){ [ 1201.518793] env[62569]: value = "task-1250787" [ 1201.518793] env[62569]: _type = "Task" [ 1201.518793] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.527470] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250787, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.028387] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250787, 'name': ReconfigVM_Task, 'duration_secs': 0.268886} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.028683] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Reconfigured VM instance instance-00000071 to attach disk [datastore1] 3e0b6e07-b175-403d-9e53-c52e34a1c507/3e0b6e07-b175-403d-9e53-c52e34a1c507.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1202.029320] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9699def6-de66-4e9c-bb1d-281c477b9b9e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.035574] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Waiting for the task: (returnval){ [ 1202.035574] env[62569]: value = "task-1250788" [ 1202.035574] env[62569]: _type = "Task" [ 1202.035574] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.043849] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250788, 'name': Rename_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.545007] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250788, 'name': Rename_Task, 'duration_secs': 0.132707} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.545349] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1202.545573] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4f0a0e58-188e-4730-9e28-73e13fbccc3b {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.551578] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Waiting for the task: (returnval){ [ 1202.551578] env[62569]: value = "task-1250789" [ 1202.551578] env[62569]: _type = "Task" [ 1202.551578] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.563180] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250789, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.062518] env[62569]: DEBUG oslo_vmware.api [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250789, 'name': PowerOnVM_Task, 'duration_secs': 0.403225} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.062811] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1203.063031] env[62569]: INFO nova.compute.manager [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Took 4.78 seconds to spawn the instance on the hypervisor. [ 1203.063223] env[62569]: DEBUG nova.compute.manager [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1203.064008] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67079603-9c43-4cd9-8dd8-20132c67c43a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.585403] env[62569]: INFO nova.compute.manager [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Took 9.47 seconds to build instance. [ 1203.957468] env[62569]: INFO nova.compute.manager [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Rebuilding instance [ 1203.999016] env[62569]: DEBUG nova.compute.manager [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1203.999905] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d124446f-14b2-4848-bfee-e60bac260606 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.087579] env[62569]: DEBUG oslo_concurrency.lockutils [None req-f102b907-dcd0-4006-a691-fec7b44d830f tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Lock "3e0b6e07-b175-403d-9e53-c52e34a1c507" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 10.977s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1205.013269] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1205.013637] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-db2d89f7-0207-4f3c-856e-83efbba011d1 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.021029] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Waiting for the task: (returnval){ [ 1205.021029] env[62569]: value = "task-1250790" [ 1205.021029] env[62569]: _type = "Task" [ 1205.021029] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.028660] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250790, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.530842] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250790, 'name': PowerOffVM_Task, 'duration_secs': 0.186109} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.531119] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1205.531357] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1205.532123] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78fae5c8-2b23-4e93-9840-f1ed77c0d9fd {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.538762] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1205.539017] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-60034b38-ae77-4621-ac59-01cba8101ed2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.561203] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1205.561431] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1205.561622] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Deleting the datastore file [datastore1] 3e0b6e07-b175-403d-9e53-c52e34a1c507 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1205.561928] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6bfc0086-8859-4872-b004-cca9a1c96cb5 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.569065] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Waiting for the task: (returnval){ [ 1205.569065] env[62569]: value = "task-1250792" [ 1205.569065] env[62569]: _type = "Task" [ 1205.569065] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.576448] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250792, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.080156] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250792, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.104354} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.080519] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1206.080617] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1206.080761] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1207.113712] env[62569]: DEBUG nova.virt.hardware [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T15:30:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T15:29:52Z,direct_url=,disk_format='vmdk',id=f77800cf-af0d-4e9c-b312-2c59488f2c7c,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='3c6a185eb8064b99b7009ae3b3fe4b5a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T15:29:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1207.114017] env[62569]: DEBUG nova.virt.hardware [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Flavor limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1207.114204] env[62569]: DEBUG nova.virt.hardware [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Image limits 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1207.114452] env[62569]: DEBUG nova.virt.hardware [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Flavor pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1207.114678] env[62569]: DEBUG nova.virt.hardware [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Image pref 0:0:0 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1207.114897] env[62569]: DEBUG nova.virt.hardware [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62569) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1207.115181] env[62569]: DEBUG nova.virt.hardware [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1207.115408] env[62569]: DEBUG nova.virt.hardware [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1207.115651] env[62569]: DEBUG nova.virt.hardware [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Got 1 possible topologies {{(pid=62569) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1207.115865] env[62569]: DEBUG nova.virt.hardware [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1207.116120] env[62569]: DEBUG nova.virt.hardware [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62569) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1207.117041] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5298cefb-cc3a-4eca-9e1d-27f04db94389 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.124681] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7d6444e-8b33-4a35-8512-27a3caccaa62 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.138050] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Instance VIF info [] {{(pid=62569) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1207.143613] env[62569]: DEBUG oslo.service.loopingcall [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1207.143847] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Creating VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1207.144074] env[62569]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-62ce95d1-af78-4f85-811d-5920c251bcdc {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.159411] env[62569]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1207.159411] env[62569]: value = "task-1250793" [ 1207.159411] env[62569]: _type = "Task" [ 1207.159411] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.167008] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250793, 'name': CreateVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.668689] env[62569]: DEBUG oslo_vmware.api [-] Task: {'id': task-1250793, 'name': CreateVM_Task, 'duration_secs': 0.229164} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.668852] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Created VM on the ESX host {{(pid=62569) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1207.669286] env[62569]: DEBUG oslo_concurrency.lockutils [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1207.669453] env[62569]: DEBUG oslo_concurrency.lockutils [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1207.669785] env[62569]: DEBUG oslo_concurrency.lockutils [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1207.670047] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc2ca3c6-bf03-4664-825f-912270bc55b2 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.674348] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Waiting for the task: (returnval){ [ 1207.674348] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52174e16-9ab9-32e8-4e56-1b77cec3fa4a" [ 1207.674348] env[62569]: _type = "Task" [ 1207.674348] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.681499] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52174e16-9ab9-32e8-4e56-1b77cec3fa4a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.184203] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52174e16-9ab9-32e8-4e56-1b77cec3fa4a, 'name': SearchDatastore_Task, 'duration_secs': 0.008068} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.184547] env[62569]: DEBUG oslo_concurrency.lockutils [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1208.184745] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Processing image f77800cf-af0d-4e9c-b312-2c59488f2c7c {{(pid=62569) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1208.184980] env[62569]: DEBUG oslo_concurrency.lockutils [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1208.185150] env[62569]: DEBUG oslo_concurrency.lockutils [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1208.185342] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1208.185631] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d26905e3-72e8-4f98-96fd-d39844ba4335 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.193403] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62569) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1208.193587] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62569) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1208.194289] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9df3b8f-69dd-4d3c-83c6-b142e352efa3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.199121] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Waiting for the task: (returnval){ [ 1208.199121] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527c13d2-cce4-3e6c-83f8-ee75de0ca6c5" [ 1208.199121] env[62569]: _type = "Task" [ 1208.199121] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.206557] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527c13d2-cce4-3e6c-83f8-ee75de0ca6c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.708639] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]527c13d2-cce4-3e6c-83f8-ee75de0ca6c5, 'name': SearchDatastore_Task, 'duration_secs': 0.008315} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.709487] env[62569]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62c3ce65-7631-4cee-af4d-4383ebc80ee3 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.714276] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Waiting for the task: (returnval){ [ 1208.714276] env[62569]: value = "session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52cc6e97-3ab5-acfa-6eaf-375de213cb8d" [ 1208.714276] env[62569]: _type = "Task" [ 1208.714276] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.721380] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52cc6e97-3ab5-acfa-6eaf-375de213cb8d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.224500] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': session[52e53cd6-f09e-d8d1-2dd0-1b3564dc6de4]52cc6e97-3ab5-acfa-6eaf-375de213cb8d, 'name': SearchDatastore_Task, 'duration_secs': 0.008668} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.224898] env[62569]: DEBUG oslo_concurrency.lockutils [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1209.225032] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 3e0b6e07-b175-403d-9e53-c52e34a1c507/3e0b6e07-b175-403d-9e53-c52e34a1c507.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1209.225291] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b7b690da-de95-4fc5-b284-11112a69ea50 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.231683] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Waiting for the task: (returnval){ [ 1209.231683] env[62569]: value = "task-1250794" [ 1209.231683] env[62569]: _type = "Task" [ 1209.231683] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.238755] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250794, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.741775] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250794, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.40506} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.742044] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f77800cf-af0d-4e9c-b312-2c59488f2c7c/f77800cf-af0d-4e9c-b312-2c59488f2c7c.vmdk to [datastore1] 3e0b6e07-b175-403d-9e53-c52e34a1c507/3e0b6e07-b175-403d-9e53-c52e34a1c507.vmdk {{(pid=62569) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1209.742263] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Extending root virtual disk to 1048576 {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1209.742514] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1ca4d14d-b6b2-46f6-ac05-1f62b72d9792 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.749585] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Waiting for the task: (returnval){ [ 1209.749585] env[62569]: value = "task-1250795" [ 1209.749585] env[62569]: _type = "Task" [ 1209.749585] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.756940] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250795, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.259187] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250795, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.24085} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.259557] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Extended root virtual disk {{(pid=62569) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1210.260199] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34bd6c37-5af6-40f9-8b36-d9f55ad8327c {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.279383] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Reconfiguring VM instance instance-00000071 to attach disk [datastore1] 3e0b6e07-b175-403d-9e53-c52e34a1c507/3e0b6e07-b175-403d-9e53-c52e34a1c507.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1210.279625] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33b4f8fb-7d7c-4478-99b7-6ea7ec8b1682 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.297723] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Waiting for the task: (returnval){ [ 1210.297723] env[62569]: value = "task-1250796" [ 1210.297723] env[62569]: _type = "Task" [ 1210.297723] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.304901] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250796, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.807076] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250796, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.308143] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250796, 'name': ReconfigVM_Task, 'duration_secs': 0.941944} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.308528] env[62569]: DEBUG nova.virt.vmwareapi.volumeops [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Reconfigured VM instance instance-00000071 to attach disk [datastore1] 3e0b6e07-b175-403d-9e53-c52e34a1c507/3e0b6e07-b175-403d-9e53-c52e34a1c507.vmdk or device None with type sparse {{(pid=62569) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1211.308980] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-62c944e6-0eba-4af6-a456-c61d7e3db43a {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.314041] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Waiting for the task: (returnval){ [ 1211.314041] env[62569]: value = "task-1250797" [ 1211.314041] env[62569]: _type = "Task" [ 1211.314041] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.321441] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250797, 'name': Rename_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.823545] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250797, 'name': Rename_Task, 'duration_secs': 0.127279} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.823855] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Powering on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1211.824111] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8598285f-2c1a-47a9-87be-8423862ccf74 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.830196] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Waiting for the task: (returnval){ [ 1211.830196] env[62569]: value = "task-1250798" [ 1211.830196] env[62569]: _type = "Task" [ 1211.830196] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.837173] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250798, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.339991] env[62569]: DEBUG oslo_vmware.api [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250798, 'name': PowerOnVM_Task, 'duration_secs': 0.38868} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.340331] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Powered on the VM {{(pid=62569) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1212.340505] env[62569]: DEBUG nova.compute.manager [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Checking state {{(pid=62569) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1212.341286] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc7d7a42-3bed-446b-be76-a94a28c4248e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.859046] env[62569]: DEBUG oslo_concurrency.lockutils [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.859292] env[62569]: DEBUG oslo_concurrency.lockutils [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1212.859477] env[62569]: DEBUG nova.objects.instance [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62569) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1213.868306] env[62569]: DEBUG oslo_concurrency.lockutils [None req-982da6aa-70ed-4835-bcba-3b74ba6ed73e tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1213.887451] env[62569]: DEBUG oslo_concurrency.lockutils [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Acquiring lock "3e0b6e07-b175-403d-9e53-c52e34a1c507" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1213.887693] env[62569]: DEBUG oslo_concurrency.lockutils [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Lock "3e0b6e07-b175-403d-9e53-c52e34a1c507" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.887901] env[62569]: DEBUG oslo_concurrency.lockutils [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Acquiring lock "3e0b6e07-b175-403d-9e53-c52e34a1c507-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1213.888106] env[62569]: DEBUG oslo_concurrency.lockutils [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Lock "3e0b6e07-b175-403d-9e53-c52e34a1c507-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.888282] env[62569]: DEBUG oslo_concurrency.lockutils [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Lock "3e0b6e07-b175-403d-9e53-c52e34a1c507-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1213.890515] env[62569]: INFO nova.compute.manager [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Terminating instance [ 1214.394856] env[62569]: DEBUG oslo_concurrency.lockutils [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Acquiring lock "refresh_cache-3e0b6e07-b175-403d-9e53-c52e34a1c507" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1214.395064] env[62569]: DEBUG oslo_concurrency.lockutils [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Acquired lock "refresh_cache-3e0b6e07-b175-403d-9e53-c52e34a1c507" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1214.395244] env[62569]: DEBUG nova.network.neutron [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Building network info cache for instance {{(pid=62569) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1214.915214] env[62569]: DEBUG nova.network.neutron [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1214.973131] env[62569]: DEBUG nova.network.neutron [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1215.475954] env[62569]: DEBUG oslo_concurrency.lockutils [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Releasing lock "refresh_cache-3e0b6e07-b175-403d-9e53-c52e34a1c507" {{(pid=62569) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1215.476425] env[62569]: DEBUG nova.compute.manager [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Start destroying the instance on the hypervisor. {{(pid=62569) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1215.476630] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Destroying instance {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1215.477920] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46f7c900-684e-40f2-8851-753a089b395d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.485533] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Powering off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1215.485787] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6ecb38f8-dca7-45f9-b875-07166961b1df {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.491926] env[62569]: DEBUG oslo_vmware.api [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Waiting for the task: (returnval){ [ 1215.491926] env[62569]: value = "task-1250799" [ 1215.491926] env[62569]: _type = "Task" [ 1215.491926] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.499766] env[62569]: DEBUG oslo_vmware.api [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250799, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.002500] env[62569]: DEBUG oslo_vmware.api [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250799, 'name': PowerOffVM_Task, 'duration_secs': 0.181891} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.002878] env[62569]: DEBUG nova.virt.vmwareapi.vm_util [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Powered off the VM {{(pid=62569) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1216.002878] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Unregistering the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1216.003105] env[62569]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-047b90f4-4984-428e-88fe-e50727eaf47d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.029511] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Unregistered the VM {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1216.029756] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Deleting contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1216.029936] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Deleting the datastore file [datastore1] 3e0b6e07-b175-403d-9e53-c52e34a1c507 {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1216.030223] env[62569]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-18e76f96-8290-467f-8bb2-b77bcda54159 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.037307] env[62569]: DEBUG oslo_vmware.api [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Waiting for the task: (returnval){ [ 1216.037307] env[62569]: value = "task-1250801" [ 1216.037307] env[62569]: _type = "Task" [ 1216.037307] env[62569]: } to complete. {{(pid=62569) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.044549] env[62569]: DEBUG oslo_vmware.api [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250801, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.547012] env[62569]: DEBUG oslo_vmware.api [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Task: {'id': task-1250801, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.088042} completed successfully. {{(pid=62569) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.547288] env[62569]: DEBUG nova.virt.vmwareapi.ds_util [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Deleted the datastore file {{(pid=62569) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1216.547479] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Deleted contents of the VM from datastore datastore1 {{(pid=62569) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1216.547660] env[62569]: DEBUG nova.virt.vmwareapi.vmops [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Instance destroyed {{(pid=62569) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1216.547905] env[62569]: INFO nova.compute.manager [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Took 1.07 seconds to destroy the instance on the hypervisor. [ 1216.548192] env[62569]: DEBUG oslo.service.loopingcall [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62569) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1216.548394] env[62569]: DEBUG nova.compute.manager [-] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Deallocating network for instance {{(pid=62569) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1216.548489] env[62569]: DEBUG nova.network.neutron [-] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] deallocate_for_instance() {{(pid=62569) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1216.563189] env[62569]: DEBUG nova.network.neutron [-] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Instance cache missing network info. {{(pid=62569) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1217.066052] env[62569]: DEBUG nova.network.neutron [-] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Updating instance_info_cache with network_info: [] {{(pid=62569) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1217.568554] env[62569]: INFO nova.compute.manager [-] [instance: 3e0b6e07-b175-403d-9e53-c52e34a1c507] Took 1.02 seconds to deallocate network for instance. [ 1218.076537] env[62569]: DEBUG oslo_concurrency.lockutils [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1218.076971] env[62569]: DEBUG oslo_concurrency.lockutils [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.077097] env[62569]: DEBUG nova.objects.instance [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Lazy-loading 'resources' on Instance uuid 3e0b6e07-b175-403d-9e53-c52e34a1c507 {{(pid=62569) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1218.613803] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bbefc2c-f3c2-4a8d-b499-5e80dbee1b1d {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.621174] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94201a8a-5516-4fe0-abb5-cf6395b0c2f0 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.650746] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-923d6fbd-7f0b-4175-90ef-724a010d5c09 {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.657504] env[62569]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beceb561-a4a5-456b-a7d2-0d489d73538e {{(pid=62569) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.669892] env[62569]: DEBUG nova.compute.provider_tree [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Inventory has not changed in ProviderTree for provider: fa06556a-5785-4014-b8bd-bc240a0cf716 {{(pid=62569) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1219.173144] env[62569]: DEBUG nova.scheduler.client.report [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Inventory has not changed for provider fa06556a-5785-4014-b8bd-bc240a0cf716 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62569) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1219.678933] env[62569]: DEBUG oslo_concurrency.lockutils [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.602s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1219.710834] env[62569]: INFO nova.scheduler.client.report [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Deleted allocations for instance 3e0b6e07-b175-403d-9e53-c52e34a1c507 [ 1220.218472] env[62569]: DEBUG oslo_concurrency.lockutils [None req-27c3ce16-fccf-4945-a46d-5ba6995eab2d tempest-ServerShowV254Test-41818761 tempest-ServerShowV254Test-41818761-project-member] Lock "3e0b6e07-b175-403d-9e53-c52e34a1c507" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.331s {{(pid=62569) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1236.156598] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1238.656209] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1240.656810] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1240.657272] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1240.657272] env[62569]: DEBUG oslo_service.periodic_task [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62569) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1240.657421] env[62569]: DEBUG nova.compute.manager [None req-20be367a-af7b-4373-82c8-62a2759c10e9 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62569) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10927}}